Publish.MinistryOfInternet.eu

Reader

Read the latest public posts from Publish.MinistryOfInternet.eu.

from Explorations of probabilistic thinking

This blog goes out to people who are not professional statisticians or data analysts, but who need to work with one. Or hire one. If this describes you, here’s something very important: you need to know about archetypes of statistical practice.

In particular, if you need support for a problem that does not directly involve clinical trials, you should be discerning of what skills you need, and you can’t assume that just any statistician will do.

So, what in the world is an archetype of statistical practice?

In my nearly-30-year career as a statistician working in industry, it’s struck me that there are three main patterns of statistical work: – The clinical statistician – The industrial statistician – The machine-learning or algorithmic statistician

These are work areas, which call for certain work practices or mental orientations. Any specific statistician naturally adopts the practice pattern for their area of work, or they naturally migrate to the work that fits their orientation pattern. I call these “archetypes” of statistical orientation.

A person can have more than one archetype, and some may be able to adapt their work pattern to the different work at hand. But some strongly live in one area predominately, and would perform poorly in a different context, if they don’t receive specific training.

You may be asking, “Where do data scientists fit into this system?” I first came up with this system long ago, before “data scientist” was a thing, so humor me and let me parse the statistical world first, and then I’ll make an attempt at placing data scientists into it.

A caveat: my career experience has been in the health field (diagnostic devices and, a bit, pharmaceuticals). If you’re coming from a different background, perhaps things appear a bit different to you? I’d be interested to hear your thoughts.

With that, the three archetypes follow.

The clinical statistician

The clinical statistician supports clinical studies interpreted in an inherently adversarial context, whether they’re supporting a research manuscript for publication or a drug application to health authorities. They make arguments concerning the appropriate degree of evidence contained in the data for or against a hypothesis, sufficient to convince a skeptical audience. Therefore they must make sure everything is in order and all critical assumptions are met.

“Biostatistics” is the subfield of statistics that focuses on clinical trials. There are challenges that arise uniquely in clinical trials, and general statisticians may not be that familiar with some of them.

To illustrate a situation in which argument is key, and it’s critical to not make mistakes: in graduate school I took a math course called “Real Analysis”, which is the theoretical development of calculus. For one exam we were given one hour to prove 10 statements. In that hour I finished proving three of them; for the others I wrote down some thoughts but did not fully prove the statements. I was careful to say what I was confident in and why.

For confidently proving only 30% of the statements, I earned an “A” on the exam. I demonstrated that I was not extremely creative in math, but I knew what followed from what, and the latter is a valuable skill. In fact, mathematics is an edifice in which one true statement is based on another, so it strongly depends on there being no wrong things in the framework. This carries over to arguing for a skeptical audience: say what you can, don’t say what you can’t, and above all, don’t say a wrong thing. A wrong thing could cause your whole case to crumble. And if they find you saying one wrong thing, and you’re unaware that it’s wrong, can they believe other things you’re saying?

I’ve seen clinical statisticians look at a careful post-hoc exploratory analysis (which can be critical to the business) and make a blanket statement that statistical inference in exploratory analyses can’t be relied upon. This is true, but quite unhelpful–more helpful would be: how compromised is the particular exploratory analysis? This is one example where an orientation is not wrong but inappropriate to the context.

The industrial statistician

The industrial statistician focuses on experiments with many controllable factors, such as in a lab experiment. Their goal is to optimize a product or process, or to develop a predictive model of a system. Here it is critical to determine the most important contributors to the process and to understand different sources of variability. Once the most important factors are identified and characterized, less-important factors are of little interest. Assessing evidence for relationships sufficient to convince a skeptical audience is not important; what is important instead is reaching findings that will move the project forward usefully, even if they’re imperfect approximations of reality.

There is a technology for handling a large number of predictor variables under experimental control, especially where every single observation is expensive (e.g., it requires an entire run of a pilot manufacturing line).

For instance, the field of “fractional factorial designs” can develop a useful local model for 5 controlled factors by collecting only 9 runs (8 runs that perturb all 5 factors and one run in the nominal “center” in order to detect whether a linear model is adequate); an additional 10th run at the center point again would be valuable in order to estimate pure error. This assumes that interactions involving 3 factors or more are trivial, which is realistic in most cases. 9 or 10 runs stands in contrast to the 25 = 32 runs one might naïvely expect as a minimal perturbation of 5 factors (not counting center points or replication). 32 runs will support estimation of all possible interactions up to 5-way, but it is highly, highly unlikely that all such interactions will be active. Finding a balanced subset of runs in 5-dimensional space is a nontrivial exercise, and a geometrically appealing one. This is a useful skill, and one that a clinical statistician could very well have no awareness of.

There is an analogy with numerical optimization routines. Note that such algorithms work iteratively, and at each iteration they make a local linear or quadratic approximation to the function of interest. It isn’t critical that the function be truly linear or quadratic; the approximation only needs to be good enough to move the search process forward. The same is true for modeling of experiment data: the model needs to be a good enough local approximation that it moves the project forward; it need not be correct.

In fact, industrial work is usually iterative, in its best form. In fact, a rule of thumb is to spend no more than 20% of your budget on your first experiment, because you fully expect subsequent experiments. Subsequent experiments can incorporate previous findings, and it is efficient not to waste resources on factors or estimates that, given prior data, can be neglected. In a sequential context, correctness is less important than in the clinical context, provided results are correct enough; after all, if the conclusions of one experiment are a little off, the next experiment will refine them.

There is an art to working with a team to elicit a list of all potential factors, develop a strategy to handle them, and to bring the team on board with the experimental design. Thus there is a bit that is ineffable and social in the practice of industrial statistics.

The machine-learning or algorithmic statistician

Statistics is about learning from data, and standard statistical methods develop this learning by making assumptions that might be more useful than true. Do errors follow a Gaussian (normal) distribution? Are relationships linear? And we begin to realize that these assumptions, while convenient, are not actually required or even the main idea. We can do about as well if we assume that a relationship is smooth, rather than linear, for instance. This direction of interest leads to semiparametric modeling, multivariate clustering, semiparametric density estimation, and predictive models that allow for arbitrary complexity of interaction (neural nets, random forest, support vector machines). This flavor of statistician also takes responsibility for assessing a model’s generalizability to future data.

However, the big dichotomy in statistics is between clinical and industrial; the machine-learning orientation is usually added to one of the others, when it is expressed at all.

Data scientists and statisticians

Where is the boundary between data science and statistics? Is there one? Much ink has been spilled on this so it is probably foolish to pursue it here…but what the heck, completeness demands it. My own experience (with clinical biomarker exploration) suggests the following: – Data scientists are intentional about the craft of programming and managing data. While a few statisticians have intentionally nurtured their craft, the community doesn’t see it as a universal need. A randomly-selected data scientist is likely to be a better programmer than a randomly-selected statistician. – Data scientists take ownership of data pipelines and data handling, more so than statisticians. – Statisticians own the question of inference. If you’re making statistical inference, you’re doing statistics. Whether or not you consider yourself a statistician. – Data scientists tend to take more responsibility than statisticians for understanding the scientific background. – I see a bifurcation in the data science community: there are those whose analysis process is to use loops and a small set of hypothesis tests and plots, then interpret the pile of results that result. Others adopt machine-learning and clustering methods eagerly.

What does this all mean?

In the health-related industry, most statisticians work in the clinical area, and support the clinical archetype. This is very important work with its own idiosyncracies, and it’s good that these practitioners generally adopt the archetype appropriate to the work. Some of these practitioners may also be able to adopt one or more of the other archetypes when placed into a different context, while others may not.

The proportion of clinical statisticians among statisticians in health-related areas is so high that many in the industry don’t realize that there is anything else. They refer to all statisticians as “biostatisticians” and expect that clinical statisticians can address any statistical need. This is a big error and can lead to business and project issues.

It would be just as big of an error to place one of the other archetypes into a clinical role if they cannot adopt the clinical orientation. However, given that the clinical role is so prevalent and highly developed, this direction of error rarely happens, or at least it is caught right away. It is the clinical-to-other direction that is more likely to be undetected, and to lead to problems. All it takes is a blasé pointy-haired boss to put the wrong person into place and the stage for mayhem is set.

In essence, managers in health-related organizations need to know that not all statisticians are alike.

 
Read more...

from Cleanzen Chicago Cleaning Services

At Cleanzen Chicago Cleaning Services, we understand how busy life can get. That’s why we’ve designed a platform to connect you with skilled independent cleaners who can take care of your home. Booking your cleaning takes just two minutes, and you’ll receive confirmation within an hour. We proudly serve Chicago and the neighboring suburbs, providing reliable and thorough cleaning services. With our satisfaction guarantee, you can feel confident knowing we’ll re-clean or refund your payment if you’re not happy. Let Cleanzen Chicago Cleaning Services handle your cleaning needs so you can enjoy more free time. Book today!

Hours of operation Mon – Sun 8AM to 10PM

Payment Types Accepted Credit Card, American Express (Amex), Visa, Mastercard, Discover

Services and Products Offered Regular Cleaning, Deep Cleaning, Move In/Out Cleaning, Post Renovation, Commercial Office Cleaning, Maid Services, Green Eco-Friendly Cleaning

Areas Served Chicago, Elgin, Wheeling, Aurora, Skokie, Niles, Glenview, Vernon Hills, Wilmette, Deerfield, Morton Grove, Buffalo Grove, Hoffman Estates, Naperville, Elmhurst, Algonquin, Lake Forest, Wicker Park, Hyde Park, Streeterville, Rogers Park, Albany Park, Northfield, Mundelein, Harwood Heights, Winnetka, Northbrook, Des Plaines, Glencoe, Prospect Heights, Mount Prospect, Rosemont, Barrington, Elk Grove Village, Oak Park, Lombard, Lake Zurich, Ukrainian Village, Logan Square, Lake View, South Loop, Jefferson Park, West Town, Gold Coast, Chicago Loop, Chicago West Loop, Chicago Lincoln Park, Evanston, Palatine, Arlington Heights, Schaumburg, Highland Park, Seattle Painters, Park Ridge, Norridge, Lincolnwood, Hinsdale, La Grange, West Dundee, Uptown, River North, Andersonville, Roscoe Village, Humboldt Park, Greektown, Old Town

Contact Us: Cleanzen Chicago Cleaning Services Address: 444 N Wabash Ave #500 Chicago, IL 60611 Phone: (312) 561-3420 Email: support@cleanzen.com Website: https://cleanzen.com/chicago Map: https://www.google.com/maps?cid=6306111083136810780

Social Links: https://www.facebook.com/people/Cleanzen-Chicago-Cleaning-Services/100063756235306/ https://www.yelp.com/biz/cleanzen-chicago-cleaning-services-chicago-3 https://twitter.com/CleanzenChicago https://foursquare.com/cleanzen5117063 https://en.gravatar.com/cleanzencleaningserviceschicagoil https://www.instagram.com/cleanzenchicago

 
Read more...

from Explorations of probabilistic thinking

Jim Garrett

  1. Introduction
  2. Why binning is so seductive
  3. Statistical efficiency
  4. Representing Nature

Introduction

Over and over I’ve found my data analysis strategies to be contrary to those of many of my peers. One area of differing inclination is that, when given continuous data, I try very hard to analyze it as continuous, rather than binning it. My preference is, according to standard principles, the theoretically preferred approach: binning is discarding information and loss of information must lose statistical efficiency. Nevertheless, this textbook advice is almost universally ignored.

I’m going to argue here that binning is a bad practice, but not for statistical efficiency reasons. Instead, I argue for trying hard to analyze continuous data on the grounds of clarity. The more we meet Nature where she is, the more clearly we can understand her and make reasonable decisions.

Note: I’ve realized that this blogging platform doesn’t allow one to include figures, unless one has access to the server, which I don’t. I’m in the process of setting up my own web site, at which point I’ll move my blog there. In the meantime, in lieu of figures, I’ll include code to produce the figures. Perhaps this will offer a little tutorial benefit. At least, I’m trying to make lemonade out of a lemon.

Here’s some setup code:

library(rms)
library(mgcv)

## A crude hand-sketched example
template <-
    data.frame(x = c(0, 4, 5, 6, 7, 8,    9,  10),
	       y = c(0, 0, 1, 3, 4, 4.25, 4.5, 4.75))

## A function to linearly interpolate
tempFun <- 
    approxfun(x = template$x,
	      y = template$y)

## Generate 
set.seed(123)
datbin <- data.frame(x = runif(300, 0, 10))

## Transform template to probability scale, ranging from 0.1 to 0.8
invLogit <- function(r) 1 - 1 / (1 + exp(r))

LogitLow <- log(0.1 / 0.9) # -2.197225
LogitHigh <- log(0.8 / 0.2) # 1.386294

datbin$p <-
    invLogit((LogitHigh - LogitLow) / tempFun(10) * tempFun(datbin$x) + LogitLow)

MedianBin <- median(datbin$x) # 4.778207

datbin$XGp <-
    factor(ifelse(datbin$x >= MedianBin, "High", "Low"),
	   levels = c("Low", "High"))

set.seed(456)
datbin$YBin <- rbinom(nrow(datbin), size = 1, prob = datbin$p)

Why binning is so seductive

As mentioned, virtually any academic statistician will argue against binning categorical data on efficiency grounds, and yet the practice is pervasive. Why? I don’t know for sure, but my guesses are as follows.

First, our customers ask for it. Very often a statistician’s customer is a medical clinical researcher, and many clinical researchers are accustomed to seeing a comparison of discrete groups. In fact, many clinical researchers aren’t even aware that any analysis other than comparing groups is even possible, if the response is not continuous. For instance, if the response variable is binomial with a hypothesized probability p that depends on independent study factors, a non-statistician researcher may imagine that a probability is a property of a group, i.e., a proportion of the whole. To evaluate a probability of an outcome given a study factor, we must define a subgroup based on the factor and then count successful outcomes per subgroup. Modelers know that logistic regression can describe the probability p as changing continuously as a function of other factors: it’s possible for every single study subject to have their own distinct probability p.

Therefore a clinical researcher asks for a subgroup definition because they prefer to think that way or they’re unaware of alternatives. Next, the pliable statistical analyst may want to please his customer by fulfilling the request as quickly and directly as possible. This raises questions about what a statistical analyst’s ideal role in collaboration is, but that’s for another day….

Second, binning can simplify analysis. Continuous data cannot be depended upon to be nicely normally-distributed as in textbook cases. There can be non-normal distributions, skewness, outliers, etc. When relating to other variables, we similarly cannot rely on relationships being linear. All of these problems simply go away when we bin. Deciding not to bin is not a simple choice, but is rather a commitment to cultivate an entirely new toolbox of analysis strategies suitable for messy cases.

Given the analyst caught between customers demanding binning on the one hand, and maybe not fully confident about their continuous-data toolbox on the other hand, perhaps it’s no surprise that most analysts simply provide the customer with what the ask for.

Still, I challenge statistical analysts to start nurturing that toolbox. If you want to know how, watch this space.

Statistical efficiency

Let’s consider the question of statistical efficiency and power with a simple hypothetical example. Suppose we are assessing a possible biomarker in a clinical trial, and the biomarker is assayed on 300 subjects. Biomarker values are uniformly distributed over a range. Suppose the clinical outcome is binary. An increased biomarker value contributes to an increased probability of clinical response, and the true relationship is per the piecewise-linear curve generated as follows:

plot(0:10,
     invLogit((LogitHigh - LogitLow) / tempFun(10) * tempFun(0:10) + LogitLow),
     type = "l",
     ylim = c(0, 1),
     main = "True probability of response",
     xlab = "Biomarker value",
     ylab = "Probability of response")

(This produces a piecewise-linear curve that is somewhat sigmoidal, starting from a probability of 0.1 (where it remains for some time), and then it increases rapidly to roughly 0.7. Then it continues increasing, but at a slower rate. Over the range of the curve (from 0 to 10), it reaches a maximum probability of 0.8.)

Even though the curve is crudely piecewise linear, it has some features that complicate real-world analysis:

  • It is not linear, not even on logistic scale.
  • The probability of clinical response does not reach zero at the low end of the biomarker range, nor does it reach 1.0 at the high end. The biomarker is informative but there are clinical exceptions.
  • At the high end, increasing biomarker value contributes to increased probability, but at a slower rate.

We want to assess whether the biomarker is worth investigating further, and if so, the nature of the relationship. I’ll carry out three alternative strategies:

  1. Calculate the median value of the biomarker and split cases into biomarker “High” and “Low” groups accordingly. Apply Fisher’s Exact Test to test for association between binary biomarker group and binary clinical outcome.
  2. Fit a generalized linear model relating binary outcome to a smoothing spline estimate of the biomarker’s contribution. Use an approximate test against the null model of no relationship. Optionally, we can assess evidence for non-linearity. Obtain an estimate of the relating curve, with pointwise confidence intervals.
  3. Similarly, fit a logistic regression model, but use a natural spline expansion for continuous biomarker values. Obtain a likelihood-ratio test against the null of no relationship, and as with the GAM, assess the evidence for non-linearity and obtain an estimate of the curve mediating the relationship.

Should we include fitting a continuous model that assumes linearity? I think not, because we don’t know if that’s the case, and a non-linear relationship is quite plausible.

With alternative (1), Fisher’s Exact Test gives a p-value “< 2.2e-16”. Statistical efficiency is not an issue here.

(The following code evaluates Fisher’s Exact Test with the X grouping against the binary outcome.)

fisher.test(datbin$XGp, datbin$YBin)

With the GAM–alternative (2)–we also find <2e-16 as a p-value against the null hypothesis of no relationship. The following code generates an estimate of this relationship:

GamMod.init <- gam(YBin ~ s(x), family = binomial, data = datbin)

summary(GamMod.init)

plot(GamMod.init,
     trans = invLogit,
     ylim = c(0, 1),
     main = "Logit outcome vs biomarker",
     xlab = "Biomarker",
     ylab = "Logit of outcome probability")

(The figure shows a figure that tracks the true curve, but it is implausibly “wiggly”.)

This tracks the true curve but is too wiggly to be plausible. Manually forcing the smoothing parameter to be large enough so that the curve is almost monotonic, we find:

GamMod <- gam(YBin ~ s(x, sp = 0.05), family = binomial, data = datbin)

plot(GamMod,
     trans = invLogit,
     main = "Logit outcome vs biomarker",
     xlab = "Biomarker",
     ylab = "Logit of outcome probability")

(The code generates a less-wiggly curve that still tracks pretty well.)

The actual predicted probabilities are not very different between these estimates, actually, and they both give p-values of <2e-16 against the null model.

## initial model
summary(GamMod.init)

## Add a linear component to the initial model so that linear and non-linear 
## components can be assessed separately.
GamMod.init.lin <- gam(YBin ~ x + s(x), family = binomial, data = datbin)

summary(GamMod.init.lin)

## smoother model
summary(GamMod)

(The initial and the smoother GAM models both show a p-value for the relationship betweeen X and outcome of <2e-16. Additionally, a linear term is added to the initial model in order to assess the non-linear contribution separately. This shows that p-values for the linear and the non-linear contributions are both very small.)

Applying alternative (3), unpenalized spline regression, we obtain the following curve:

## Use rms package to enable nice ANOVA
RegModBin <- lrm(YBin ~ rcs(x, parms = 5), data = datbin)


## Use base or "stock" glm to support likelihood ratio test via base 
## anova function

RegModBin.s <- glm(YBin ~ rcs(x, parms = 5), data = datbin)

## Plot
## Set of points on X axis for plotting
TmpSeq <- seq(0, 10, length = 200)

## Get predictions on logistic scale, then calculate confidence limits
## on that scale, then transform to probability scale
Preds <-
    predict(RegModBin,
	    newdata = data.frame(x = TmpSeq),
	    type = "lp", se.fit = T)
## has components "linear.predictors" "se.fit"

CIReg <- 
    data.frame(Est = invLogit(Preds$linear.predictors),
	       Low = invLogit(qnorm(0.025,
				    mean = Preds$linear.predictors,
				    sd = Preds$se.fit)),
	       High = invLogit(qnorm(0.975,
				     mean = Preds$linear.predictors,
				     sd = Preds$se.fit)))

plot(range(TmpSeq), c(0, 1), type = "n",
     main = "Outcome probability vs. biomarker",
     xlab = "Biomarker",
     ylab = "Outcome probability")
polygon(x = c(TmpSeq, rev(TmpSeq), TmpSeq[1]),
	y = c(CIReg$Low, rev(CIReg$High), CIReg$Low[1]),
	col = "thistle", border = NA)
lines(TmpSeq, CIReg$Est)
rug(datbin$x[datbin$YBin == 0])
rug(datbin$x[datbin$YBin == 1], side = 3)

In this plot the Y-axis is on the probability scale rather than the logit scale. This is substantially equivalent to either GAM model. Here we can give an informative ANOVA breakdown:

## Generate the ANOVA table
anova(RegModBin)

Factor     Chi-Square d.f. P     
x          69.85      4    <.0001
 Nonlinear  9.89      3    0.0195
TOTAL      69.85      4    <.0001

## Likelihood ratio test using base R
anova(RegModBin.s, glm(YBin ~ 1, data = datbin))

This indicates that there is overwhelming evidence that the biomarker influences the outcome, and furthermore there is strong evidence of a nonlinear component, i.e., a departure from linearity on the logit scale. While this representation doesn’t show exactly how small the p-value is, a standard likelihood-ratio test yields <2e-16, just as the other methods.

In summary, all three approaches indicate strong evidence that the biomarker influences the clinical response. This does not support the idea that the continuous approach is more powerful. The median split represents the biomarker with 1 degree of freedom, while the continuous approaches use roughly 4 degrees of freedom. They yield more information, but “cost” more. It’s a fair trade, but it’s not clear that one always has more power than the other.

Rather, the reason that I recommend a continuous approach is that, in one step, we (1) assess evidence for a non-null relationship and (2) gain a reasonable estimate of that relationship. Further, we do the latter without carrying out substantial optimization or multiple looks at the response, which compromises statistical reliability.

Now let’s think a little more about real life.

Representing Nature

Here’s a true story illustrating how failure to look at continuous data can lead to self-imposed confusion and obfuscation, also cost real money, and delay important projects.

As the resident expert in clinical diagnostic assays in a large pharmaceutical company (that’s not saying a great deal when the company didn’t nurture such expertise), I was pulled into an apparent assay issue impinging on an oncology clinical trial. The assay measured gene copy number (GCN) for a specific gene; a GCN value above a specified cutoff was a study enrollment criterion. That is, it was a companion diagnostic assay (CDx) for the therapy under study. Two labs carried out testing for the study, each serving a different geographic region.

Recently there had been some operational issues with the assay which had required troubleshooting; the assay vendor had confirmed the issue, put a fix in place, and, for good measure, both labs repeated operator proficiency validations. Then patient screening for the study resumed. After some time, however, the trial team noticed that the “prevalence” (incidence of GCN over cutoff) was higher at one lab than the other. It was decided to pause study enrollment once again. zNote that for a pharmaceutical company, completing trials quickly is the coin of the realm; pausing a trial was a Very Serious Matter. Meetings were held, numbers were compiled, and still bigger meetings were held. Finally this expanding process grew to include a bystander previously unaware of the entire study, i.e., me.

When I joined my first meeting, it was chaired by the head of Oncology, which, for the company organization at the time, reported to the CEO. Lots of highly-paid senior people were there; this was one of those meetings where the cash clock ticked quickly.

I was given the information that had been compiled up to then. This included assay positivity counts at each site. I asked, “Where are the GCN numbers for each of the sites? What does the GCN distribution look like at each site?”

Such information had not been compiled!

Consider for a moment what this indicates about priorities and corporate culture. The company was expending significant resources, pausing a trial and using a top executive’s time. It would have been the simplest thing to organize GCN values–they were available in the clinical database, waiting to be visited. There’s no question that the values actually measured would give a more complete representation than the processed values. Yet this value was not widely shared. If this describes your organization too, you have work to do!

As an aside, there’s another aspect to this: when you’re troubleshooting a data-related issue, investigate the data process from first information acquisition to final result before you invest time in a lot of other approaches. At what point does the data begin to look anomalous? Or, if you prefer, start at the end and work towards the beginning. The point is to be systematic and to “scan” the whole process to come to an understanding of the state of things. Looking at continuous data often means looking upstream; I also have an expensive war story about this. The error is committed again and again.

But back to our story: in not much time the team pointed me to GCN numbers and I was able to determine that the clinical cutoff for GCN was very near the median GCN value. This is not necessarily an error, but it is definitely problematic: a small shift in the distribution of measured values, such as can easily happen with many assays, will induce a substantial change in positivity rate. I fitted density estimates; they had similar shapes but one was shifted slightly higher than the other. If I shifted the higher density down by the difference in medians, the densities lined up quite well, and furthermore the apparent positivity rates closely agreed.

This difference in median was smaller than the noise in the assay, so an assay scientist wouldn’t worry too much about it, and would certainly recommend against placing the assay in a context where a trivial change (relative to assay variability) would be interpreted gravely. Red flags and warnings should have gone up when the clinical cutoff was suggested.

The team decided to carry out a paired sample study. Regulatory requirements prevented either lab from sending clinical samples to the other lab, but the assay vendor could split samples, test them, and send them to each lab. Then we could compare each lab to the vendor. Long story short, it turned out there was a small difference between the vendor and both labs, and in the same direction, but this difference was not meaningful. The trial resumed. Frankly, while there was an observed difference, when interpreted with quantitative data and an understanding of variability, there was no real issue. By looking at processed data and not turning quickly to the underlying quantities when questions arose, the team had gotten themselves in very costly tizzy.

Here’s what I’ve taken away from these experiences:

  • Model continuous data when possible. It’s probably closer to the data that Nature gave you than binned or otherwise processed data. – Transforming such data towards a Gaussian distribution is fine, in fact it’s often beneficial.
  • Analyzing data close to what Nature gave you will give you a more complete assessment. It may have some analysis complications, but it will give the decision-making team more confidence.

For more on this topic, check out Frank Harrell Jr.’s essays How to Do Bad Biomarker Research and a chapter on Information Loss.

#continuousdata #dataanalysis #statistics

 
Read more...

from W nieskończoności skryta...

Aspołeczna kawiarenka

I znowu myślę o opuszczeniu moich socjali... Znowu, bo w połowie sierpnia usunęłam swoje konto na mastodonie. Niestety tak już mam. Gdy życie daje mi za mocno w kość, to uciekam od wszystkich. We wrześniu poprosiłam o ,,przywrócenie do życia” mojego starego konta, więc jakoś tam wróciłam. A że życie nie jest fajne, a ludzie tym bardziej, to znów chciałabym się schować przed całym światem. Najlepiej tak, by mnie już nikt nie znalazł.

Nie jest niczym odkrywczym, że nie potrafię się odnaleźć wśród ludzi. A już socjalmedia to coś, czego nie da się ogarnąć w jakikolwiek sposób. A przynajmniej ja tego nie potrafię. Ostatnio natknęłam się na komentarz typu ,,to nie twój stoliczek, więc się nie wtrącaj do dyskusji” a później jeszcze, że ,,jak ktoś ma chorobę morską to nie pcha się na statek”. Nie brałam udziału w tamtej dyskusji, ale były to osoby znajome, z którymi raz na jakiś czas rozmawiam i ogólnie mam z nimi jakiś kontakt.

No więc ja już od dawna nie wędruję po ,,obcych” stoliczkach. To znaczy, nie odzywam się za bardzo pod postami innych ludzi. Ograniczam się do kilku osób, uznanych przeze mnie za ,,w jakimś stopniu bezpieczne”. Siedzę głównie przy ,,swoim” stoliczku i jak ktoś się odezwie do mnie to odpowiadam. I, wbrew temu co myśli sobie osoba, która wspomniała o unikaniu niebezpiecznych miejsc, wydaje mi się, że mój ,,własny” stoliczek też nie jest dla mnie bezpieczny. Bo co z tego, że ja unikam osób, które dla mnie stanowią zagrożenie. Korzystając z analogii do choroby morskiej. Ja mam chorobę morską, dlatego nie zapuszczam się w morze. Ale ci, co nie mają choroby morskiej i pływają sobie po tym morzu, nie mają też choroby lądowej, więc nie unikają miejsc w których ja bywam. W szczególności nie unikają mnie, bo im nie zagrażam w żaden sposób. Zazwyczaj osoby, których z jakichś powodów unikam, nie mają żadnych oporów, żeby odzywać się do mnie. A to dla mnie oznacza ryzyko. Bo nawet gdy mam w nazwie tęczową nieskończoność, to i tak większość osób na to nie zwraca uwagi albo nie mają pojęcia z czym to się wiąże. Pomijam ludzi, którym się po prostu nie chce, oprócz nich są jeszcze tacy, którzy zapominają. Spotkałam się też z osobą, która wiedząc, że jako autystka mam problem z ironią i sarkazmem ,,pozwoliła sobie na odrobinę ironii”. No bo dlaczego nie...

W cuda nie wierzę, w ludzi tym bardziej... Dlatego bezpiecznym miejscem dla mnie jest moja własna szafa. Powinnam tam sobie siedzieć, nie wychylać się i tylko kilku osobom dać namiary na tę moją szafę w nadziei, że o mnie nie zapomną i czasem się odezwą...

 
Czytaj dalej...

from W nieskończoności skryta...

Chaos

Moim nieodłącznym towarzyszem życia zawsze był chaos. Czasami mniejszy, czasami większy, ale zawsze był. I choćbym nie wiem jak bardzo się starała, to nigdy nie potrafiłam sobie odpowiednio zaplanować i poukładać ani dnia ani tygodnia. W zasadzie jedyny czas kiedy moje życie było w jakiś sposób regularne to czas szkoły, studiów i te kilka lat pracy. Jak widać, była to regularność narzucona z zewnątrz, do której potrafiłam się jakoś dostosować. Nie zawsze perfekcyjnie, ale w jakimś stopniu mi się udawało, przynajmniej jeśli chodzi o obecność. Sama z siebie jednak nie potrafiłam nigdy wprowadzić żadnej rutyny. I nie chodzi tu o brak umiejętności zaplanowania sobie dnia czy tygodnia pracy. O nie, plany to ja potrafię robić. Problem zaczyna się z ich realizacją.

Dzieje się tak, gdyż nie potrafię w żaden sposób odnaleźć się w czasie. Nie wiem dlaczego, ale nie potrafię odpowiednio zgrać czynności z czasem. Np. jak mam gdzieś być o konkretnej godzinie to albo jestem dużo za wcześnie, albo biegnę na ostatnią chwilę. Nie trudno zauważyć, że marnuję sporo czasu wiecznie gdzieś na coś czekając. Drugą rzeczą, która utrudnia mi realizację planów jest jakieś wewnętrzne coś, nie wiadomo co, które siedzi sobie we mnie i jak tylko sobie postanowię, że zrobię jakąś rzecz, to automatycznie to coś sprawia, że nie jestem w stanie zrealizować zaplanowanej rzeczy. Nie wiem co to jest i z czego wynika. Najbardziej oczywiście owo coś przeszkadzało mi przed wszelkiego rodzaju egzaminami, kiedy ważne było zaplanowanie sobie nauki.

Jak tak sobie patrzę z perspektywy czasu i analizuję swoje życie, to okazuje się, że nigdy tak naprawdę nie uczyłam się do egzaminów. Ani do matury, ani na studiach. Akurat jeśli chodzi o egzaminy typu matura i sesje na studiach, to oprócz braku umiejętności zrealizowania planu nauki/powtórek, czynnikiem utrudniającym był stres i presja czasu. Im bliżej egzaminu tym większa niemoc zrobienia czegokolwiek. Dlatego mój sukces na egzaminach nieodzownie związany był z regularnym uczęszczaniem na zajęcia i samodzielną pracą w ich trakcie a także samodzielnym wykonywaniem tzw. prac domowych czy robieniem po prostu w domu dodatkowych zadań, bo najzwyczajniej w świecie sprawiało mi to przyjemność. Jak się można domyślić, jeśli czegoś nie lubiłam to miałam problem.

Z powodu tego wewnętrznego czegoś, czasami łatwiej zrobić mi coś spontanicznego, niż zrealizować plan. Czyli wypisz, wymaluj chaos.

Pomijam już rzeczy typu, zaczynam robić jedną rzecz, a kończę zupełnie z czymś innym, bo akurat coś zobaczyłam, usłyszałam itp... Kolejny temat rzeka a raczej łańcuszek czynności, nie do końca powiązanych ze sobą.

Ale ten chaos to nie tylko trudności w takim, codziennym funkcjonowaniu. To także trudności na mojej drodze do odnalezienia siebie. Bo ten wszechobecny w moim życiu chaos sprawiał, że przez długi czas nijak nie potrafiłam odnaleźć w sobie spektrum autyzmu, o którym wcześniej wielokrotnie słyszałam i próbowałam nawet czegoś o spektrum szukać ale odbijałam się za każdym razem od rutyny, której u mnie nie mogłam po prostu dostrzec w żaden sposób.

Gdy odkryłam o sobie prawdę, chaos stał się powodem do niepokoju. Bo jak już odkryłam, to szukałam osób podobnych do mnie, między innymi w tych moich socjalmediach. Od dawna miałam jedną koleżankę, dzięki której w zasadzie powoli zaczęłam ruszać temat. Później kolejna osoba z sieci. Co mnie zaczęło niepokoić, to właśnie ich bardzo poukładane życie. Ja takiego nie mam. U mnie rutyna jest, ale na opak. Dotyczy tylko konkretnych rzeczy. A więc jak mam biegać to muszę odstawić w domu cały rytuał przygotowawczy, a jak biegnę, to stałą trasą, z telefonem w ręku, żeby widzieć jakie mam tempo i jaki puls. Gdy tylko GPS przestaje działać, to zaczynam panikować i najczęściej nie jestem w stanie biec. Ale to tylko bieg. Natomiast nie potrafię biegać regularnie np. 3-4 razy w tygodniu. Tego już za wiele. Taka rutyna jest poza moim zasięgiem. I takich przykładów jest pełno. Dana czynność wykonywana wręcz rytualnie, ale bez zachowania regularności w czasie.

Odpowiedź na to jest tylko jedna. Jestem bardziej neuronietypowa, niż mi się wydawało. Spektrum autyzmu to tylko jeden składnik. Jest jeszcze drugi, który zaburza ten pierwszy. W zasadzie one wzajemnie na siebie oddziałują, a raczej na mnie, wprowadzając niemałe zamieszanie, przez które trudno dostrzec i jedno i drugie...

Odkrycie drugiego składnika dało mi chyba upragniony spokój. Taki wewnętrzny. Bo w końcu wszystko pasuje, układa się w jakąś jedną, spójną całość. I nawet gdy nie ułożyłam swojej układanki całkowicie, to i tak jest lepiej. Bo nie czuję już, że mam jakby dwa niepasujące do siebie zestawy puzzli.

Poza tym, już sama świadomość tej całości sprawiła, że w pewnym sensie trochę łatwiej mi w takim codziennym życiu. Bo już nie kopię się sama z sobą, nie próbuję na siłę robić rzeczy wbrew sobie, bo wiem, że to i tak jest bez sensu. I najważniejsze, nie mam już tego uczucia, że zupełnie nigdzie nie pasuję.

Tutaj tym bardziej nie będę robić niczego wbrew sobie. A zatem, w zgodzie ze sobą, wbrew temu co kiedyś pisałam, nie będzie żadnego planu tego miejsca, nie będzie regularnego pisania...

Witam w moim chaosie który tu był, jest i zawsze będzie...

 
Czytaj dalej...

from Ábrete mundo.

A pesar de que la rudeza es una pieza fundamental para la construcción de roles y esquemas de relación en nuestras culturas, me parece que se habla poco de este elefante en medio del salón cuando se habla de psicología o cuando se analiza el funcionamiento de la sociedad.

La rudeza es uno de los ingredientes que suelen utilizarse en la receta de la masculinidad. Cada hombre usará de este ingrediente en mayor o menor medida pero parece difícil prescindir de él, como del ajo en las cocinas mediterráneas. El niño varón suele asumir que tiene que desarrollar su capacidad de rudeza para ser reconocido como plenamente masculino, el varón debe aprender a reprimir sus emociones, la única emoción cuya manifestación no cuestiona la masculinidad tradicional es la ira.

La rudeza es también una defensa frente a la conflictividad de las relaciones interpersonales en todos los ámbitos, una advertencia de respuesta violenta (física o verbal) ante eventuales agresiones.

La rudeza también funciona una estrategia para proteger la autoestima de la humillaciones de la convivencia y una manera de hacer algo más llevaderas las situaciones de dominación soportadas, muy especialmente en el ámbito laboral y militar, porque se supone que la rudeza revaloriza al individuo, compensando la desvalorización derivada de soportar y verse obligado a asumir situaciones de dominación. Frente a la dominación se reacciona con obediencia, que desvaloriza, y con una pose de rudeza, a la que se le atribuye un efecto valorizador, sobre todo en los machos.

El servicio militar obligatorio era una escuela de rudeza que enseñaba que el desarrollo de cierto nivel de rudeza era necesario para lograr la validación del grupo y de la autoridad. La supresión del servicio militar obligatorio ha reducido la valoración social de la rudeza, sin embargo las abundantes fantasías militares de la cultura del entretenimiento, siguen enalteciendo la rudeza, sobre todo como elemento de la construcción de la identidad masculina, pero también como elemento a valorar en ciertas identidades femeninas, de ruptura con la niña o con comportamientos de mujer tradicional.

¿La rudeza es un elemento que la empresa valora positivamente al seleccionar personal? Podría ser percibida por el seleccionador de personal como un indicador de aguante a la presión del dominio jerárquico (el trabajador rudo podría pensarse que no es tan propicio a coger una baja por ansiedad) y como un indicador de que el candidato a trabajador no cultiva otras habilidades sociales más trasversales que pueden perjudicar el control de la plantilla por el superior jerárquico al establecer lazos entre los trabajadores o al poner en entredicho, por comparación, las habilidades del mando.

El hombre de mediana edad, normalmente enfrentado a la realidad de una actividad laboral insatisfactoria, a una proyección social insuficiente y a la frustración de no haber alcanzado las promesas vitales que ofrece a los jóvenes nuestra sociedad de mercado, suele trabajar una coraza de cierta rudeza para compensar esos pequeños (muchas veces inevitables e incluso naturales) fracasos que dañan su identidad y autoestima. La rudeza puede ser un sustitutivo del estatus no logrado, pero incluso puede acompañar a la ostentación de estatus, como un adorno más. El rock viejuno, con toda una colección de grupos que lo cultivan, hace abundante uso de una imagen de rudeza de mediana edad.

Cerveza y tatuajes se han convertido en elemento de identidad de una buena parte de la población, demostración no solo de estar en la cultura contemporánea si no también de rudeza.

Siendo la rudeza un complemento importante para las personas de clase trabajadora y clase media, se entiende el potencial de la ideología rojiparda. La izquierda con origen en el 68, la izquierda más postmoderna, no se relaciona bien con esa demanda de identidad ruda. En cambio el rojipardismo si puede cultivar la imagen ruda o al menos, aceptarla.

Quizás la rudeza debería ser un elemento tenido mucho más en cuenta por la psicología y la crítica cultural, teniendo en cuenta el papel tan importante que juega al configurar el carácter y la cultura.

 
Leer más...

from Explorations of probabilistic thinking

Some thoughts on statistical modeling and data science James Garrett

Some years ago, in a large pharmaceutical company, an academic paper circulated among some of the members of the statistical staff (of which I was one). It addressed the question of how logistic regression compared to machine learning (ML) methods in predictive accuracy with clinical trial data. The authors of the paper applied logistic regression models and found equal or better predictive performance than had been reported earlier with ML methods.

The paper was greeted by the statisticians with an uncomfortably tribal quality—I think someone may have literally written, “Hooray for our side!” If that wasn't the literal statement, the discourse within the group certainly ran along that line, such was the feeling of siege that prevailed.

(Unfortunately, I cannot find that paper now. A search on this topic uncovers many papers on the relative predictive accuracy of logistic regression on clinical data. It's been a topic of some concern, apparently.)

I had been studying statistical modeling methods advocated by Dr. Frank Harrell, Jr., whom I like to call “The most respected ignored statistician in America.” He's an elected fellow of the American Statistical Association, which is as close to a Nobel Prize as the statistics world comes. He specializes in exploratory modeling of clinical data. He mined decades' worth of statistical thinking to synthesize an exploratory modeling workflow that is purported to be thorough, efficient, and likely to yield replicable results.

The funny thing about this methodology is that at any single given point in its process, it is utterly familiar to statisticians. Logistic regression here; sure, feeling right at home. Model selection criteria there; got it. Assessment of correlations among predictor variables; of course. However, when you put all the pieces together and take it from A to Z, the whole is not familiar at all. On two occasions, with two large corporations, sat with large statistics groups and heard Dr. Harrell walk through his process. I've left the room watching audience members nodding their heads and saying the good Doctor had made a lot of convincing points. But would they incorporate these ideas in their own work? “No, my clients would never let me.” That's why I say Harrell is highly respected yet mostly ignored.

One aspect that figures prominently in Harrell's workflow is inclusion of spline expansions to enable simple non-linearity for continuous predictors. In over ten years of work within the large statistical group, I had seen lots of logistic regression models, but not one included a spline expansion.

So it was interesting indeed when I looked more closely at the paper which “scored one for our side.” It didn't give a lot of detail about how the logistic regression models were fitted, but they contained spline expansions. It seemed extremely likely that the creators of the models were aware of Harrell's process. At any rate, by using splines they were out of step with typical practice.

In fact, this paper wasn't a win for “our side,” if “our side” refers to statisticians who fit models in the typical manner. It was a win for something completely different, neither ML nor typical statistical practice. It was a win for what statistical modeling could be, but rarely is. I believe my colleagues were a little quick to accept this paper as representing their practice. It suggested a middle way between traditional statistical modeling and ML, a way which can predict as well as ML, can be as flexible in many respects, and can be more informative to boot. If the data is appropriate.

I intend to write a follow-up essay soon suggesting why I think Harrell's approach works well for most clinical data sets, so stay tuned. I'll also offer some thoughts on a way to categorize statistical and ML modeling methods according to their behavioral properties, to aid in picking the right method for the data set at hand. We really shouldn't be organized in tribes at all; there is no best modeling method, there is only a method that is best at exploiting the features of a particular data set. What are those features? I'll offer my suggestions soon.

 
Read more...

from W nieskończoności skryta...

powrót

wygląda na to że chyba wróciłam z zaświatów obłędu do świata Nic jeszcze nie w pełni sprawna jeszcze nie w pełni sił wciąż przyczajona czekam na ten właściwy moment by wreszcie zacząć normalnie żyć

 
Czytaj dalej...

from Ábrete mundo.

Hace poco acabé la novela “Elantris”, de Brandon Sanderson y ya pude escuchar el podcast https://www.ivoox.com/elantris-brandon-sanderson-audios-mp3_rf_30156272_1.html sobre la novela, que tenía reservado para escuchar cuando me acabara el libro.

El relato me gustó mucho. Es de fantasía, pero el tipo de magia que describe me hacía pensar en que podía haber un sistema tecnológico detrás, que los aones eran una interface holográfica para para manejar un sistema informatizado oculto que manejaba algún tipo de infraestructura tecnológica planetaria. Incluso me pareció que el dios Jaddeth, en su misterioso trono subterraneo esperando para salir a a superficie, podía ser una inteligencia artificial disputando el control del sistema. Los seones también tenían pinta de ser una inteligencia artificial con un cuerpo energético o con una presencia proyectada en el mundo físico.

En el podcast se compara los aones con runas, pero creo que, tal y como los describían en la novela, se parecerían más a ideogramas chinos (cuyo origen se piensa que estuvo asociado a prácticas mágicas). En cambio, la escritura de Fjordell, que usan en la religión Shu-Dereth, me parece que recordarían más a las runas.

Las tramas políticas de la novela me parecen un mecanismo de relojería muy impresionante.

Durante la lectura, el personaje de Galladón me lo imaginaba con la cara de Morgan Freeman.

La traca final con que acaba la novela me parece quizás un poco exagerada, todo muy al filo, un poco como en una película de acción de Hollywood.

Quizás a los personajes les falte un poco de textura, seguramente porque “Elantris” fue la primera novela publicada de Sanderson, pero el autor no llega a caer en una dinámica de personajes (diálogos, comportamientos...) propia del cine comercial, un pecadillo muy frecuente en autores de ciencia ficción estadounidense.

 
Leer más...