On Jan 3 I had knee surgery--patellofemoral, not a total replacement. I'm doing well with that, and I get the medical staples out tomorrow.
I'm tapering off the pain medications. Either I don't need them as much, or ... whatever. There is only so much television, so much gaming, so much mindless stuff one can do before the brain rebels. Either that, or goes completely asleep. Since it's easier to exercise the brain than to wake it up from complete dormancy, I thought I would work a bit on a project; someone at work would like their data analyzed using unsupervised methods, and the laptop I'm currently using (mid 2015 MBP, Retina, 16GB) has been giving me problems since I brought it home. It's used, and I suspect it was abused a bit prior to it going to the reseller. It has been restarting during the night when no one is using it; the fans start blazing away for no apparent reason (as it's been sitting unused for hours); it occasionally freezes.
I have been jonesing for a M1 16" MBP, M1Pro (not Max), and 32 MB integrated RAM. I can find the 16 MB version locally (I can get one today from Worst Buy), so the question becomes, do I need the increase in RAM? Have I chosen an appropriate machine for what I'm wanting to do? For me to order a new M1 with these specifications will mean the the machine will take about 6 weeks to get to me. I chose the 16" because screen real estate is precious, and I can take it with me wherever I go.
A bit more information may be needed: unsupervised methods of analysis on large data sets uses algorithms to find data that is similar to one another; such data will cluster together, as in this one is positive for this attribute or attributes, but negative for these others and it is similar to this other data point. It will point out hidden structure found in the data. At work, the data the researchers generate is normally looked at via one parameter plotted against another, and the data that fits the "preconceived" criteria is selected, and the process continues with another plot of one parameter against another using the selected data, etc. I say "preconceived" because the researcher will say "I'm looking at the data that fits this criteria and ignore the rest." This manual method is highly susceptible to bias, particularly as the number of parameters grow. The instrument I normally operate will detect up to about 16 parameters, including time, and the newest instrument in the lab will detect (I think) up to 43 parameters (including time).
For the algorithms to do what they need to do, the data needs to be cleaned up (known extraneous and confounding data need to be removed from the data set. As I will be working with data generated from blood samples as applicable to particular pathologies, for instance, dead cells and cellular debris can be eliminated), the data can then be dimensionally arranged and the algorithms let loose to do their work. Dimensionally arranged, you say? Well, think of three dimensional space, with an x, y, and z directions. Add time to that model, to get you to four dimensional. The brain begins to boggle at this idea, and one really cannot imagine it clearly. Now add more dimensions, one for each thing that is being measured. Now you have multidimensional space, and the complexity of the space (its dimensionality) increases with increasing numbers of parameters being measured.
The only time that a visual is usually needed is at the end, to see what the algorithm(s) have found. Thus, it is similar to graphical work only in that both will use large amounts of data.
This is what I'll be working with.