Logbook: Mind palace
Buy Now: pay later gas is a bad idea..

National Gas Tax suspension to fuck you over later, they are looking to suspend the tax.The current tas is at 18.4 cents a gallon. Notice the wording suspend. Not stop, likely they are going to not collect it and than raise it to 36.8 cents to make it back . Basically they are just giving you a defered loan.
President Donald Trump and congressional Republicans are proposing suspending the federal gas tax amid soaring prices at the pump caused by the war with Iran.
The gas tax is currently 18 cents per gallon, and gas prices in the U.S. are about $4.50 per gallon on average.
Voters are souring on Trump’s economy ahead of the 2026 midterm elections, and high gas prices are only adding to consumer discontent.
This is a deferment of 112.4 million a day!
Trump said in the Oval Office on Monday that he would “reduce” the tax, shortly after saying in an interview with CBS News that he wants to pause the tax “for a period of time.”
“I think it’s a great idea,” he said in the CBS interview. “Yup, we’re going to take off the gas tax for a period of time, and when gas goes down, we’ll let it phase back in.”
Sure, the Administration will defer it till the next election and blame the democrats when a billion dollar bill is due back at the pump. People have no idea how much this would affect the national infrastructure system, all of the bridges, other things that would fall into disrepair due to federal funding being frozen due to this.
Trump cannot declare a gas tax holiday alone, since Congress has sole authority over taxation. But several Republican lawmakers on Monday floated suspending the gas tax, which clocks in at 18.4 cents per gallon, the same amount it’s been since 1993.
This is unlikely to pass, but this gas lay-away plan will hit the pumps when you least expect it.
Sen. Josh Hawley, R-Mo., on Monday said he would immediately introduce a bill to suspend the federal gas tax in a post to X. And Rep. Anna Paulina Luna, R-Fla., said she would be “introducing a bill in the House to suspend the federal gas tax in light of Trump’s recent remarks.”
“American families need this relief on gas prices. My office will be working directly with President Trump to ensure we deliver this win for the American people,” she said in a social media post.
This is not relief, this is Gas HELOC plan. It will come back with interest later, because the 112.4M a day gains no interest, It is just a payment plan that is put back on the national debt. The other thing is if you look at the highway Adminstration, the infrastructure money will run out just before the election and who over gets the handoff on “The roads are falling apart” will be the media screaming for months.
Anna Paulina Luna, R-Fla “American families need this relief on gas prices. My office will be working directly with President Trump to ensure we deliver this win for the American people,” she said in a social media post.
This is not relief. its like a fart with the gas prices.. By holding the fart you save current environment because it seems not to stink, but it when comes to release and if you try to force it (collect the federal gas tax deferment) it gonna stink like when you are sick and can’t hold it anymore, and you try to let a little out and you shit yourself..
Gas prices in the U.S. currently sit at about $4.52 per gallon, according to AAA’s national average, approaching the highest recorded average price of $5.02 in June 2022. Slashing the federal gas tax would bring that down to roughly $4.34 per gallon.
This is just is a credit card at the pump, you save 18 cents at the pump and by the time the next time it comes to collect the taxes you’ll have to pay 36 cents because of interest.
In the end with a 14 gallon tank you save $2.58 and you can maybe afford cheap gas station coffee, that you have to pay back later. It’s a sham if they defer costs.
While the bill has not been filed yet. this is my thought on a gas tax moratorium to delay or defer gas taxes, in the end it will effect more than saving on your gas.
Attributions from:
CNBC: Trump, congressional Republicans float suspending federal gas tax amid Iran war
From the be careful of what you wish for: Removing the VRA and the case of the bad Genie wish.. (Voting Rights Amendment)

I have been thinking about this for a bit, and removing the voting rights act by itself is a bad genie wish. Yes its removed but, The people celebrating have no idea what is about to be unleashed. Rather than a group now, The individual mandate will apply.
The fact of the situation is before a group could sue the state about being discriminated on by the state. When the SCOTUS ruled to take apart this rule they may have angered the wrong wish genie, Before this ruling, the Voting Rights Act acted as a filter. It forced people into Class Actions.. Now that they have removed this an individual can now sue the state as in individual rather than a class, Meaning whatever complaint you have in a situation if you “feel” it wrongs you now you have the ability to sue for.
Now using Intent as a standard, a plaintiff’s personal experience is the starting point. If you “feel” sorted or excluded, the court is legally obligated to investigate your claim of discrimination. The Trap: Because “intent” is subjective, the court can’t dismiss you based standing. They have to allow for discovery searching the state’s private records, to see if your “feeling” is right.
The thing is the group that wanted this gone have no idea what on earth they just released from pandora’s box. Now that groups can not sue, It will allow for sole people to sue, and sue pro se under the law, it sounds like a bigger hill to climb, but it actually creates the bad wish part, Groups of individuals can no sue pro se and courts can not dismiss these cases in one sweep. where the VRA protected the courts before, The VRA was like a dam it gathered all the voters up into one location and as a single form allowed the courts to see the issue as a singular subject. No more, voters can not all file a case, Without a lawyer, and whereas before you had (Group 1) sueing , now you can have tens of thousands sueing all at once, with Lawyers, pro se, Pro per, and so on. So if this is what the scotus wanted they are going to end up locking up every court from the local level to the SCOTUS.
The absolutely evil end of ending the VRA , is now you can sue the state for discrimination as an individual on, to white, to colored, to disabled, to You don’t like your neighbor voting because of reasons, to you have a gene that someone else doesn’t have, to much violence, to little violence, to many cars, to many bikes, in the most basic form it has opened up warfare by skirmish on every little battle. It causes a skirmish all the way up the chain if the person starts filing on the state level based on the 14th amendment. This is unholy destructive, and now the case can be lodged at any starting point (Town,City,county,State) and how SCOTUS did not have foresight for this is bloody amazing. Given the fact that SCOTUS has enforced the individual mandate now, the ones suing for (Random Voter Discrimination issue) could basically end up stripping every town/city etc of every document they ever had the word “voter” on.
In closing, If 1% of citizens per state files an individual lawsuit due to the math of Intent, The legal system will be so bound up by this that it would effectively paralyze every court in every state. because even if 30,000 suits per state, they would have to pull every judge, every legal aid, every lawyer that works inside the local and state governments to the point of shutting down other courts (Civil, Family, criminal, Enforcement). Normally, the court would try to bundle these but now due to the individual mandate, Every single person could just tell the courts “I hereby refuse class-action status; my injury is personal and based on my specific DNA/neighborhood/feelings,” and the judges would be stuck with tens of thousands of cases, tens of thousands of filings for discovery , Appeals, motions, AMICUS Briefs, Multiplies the damage hundreds of times past the original filings. So in short, Those Judges might want to get their strongest coffee out and make a cup the size of Cleveland. Since Nationwide Injunctions was ended by the SCOTUS, they have created a monster they never had the preemption to notice. The rights of the local courts have lost the power of injunctions, they have lost the power of the stay.. it’s basically no holds barred the way I see it .
Now I am not a lawyer, I just analysed an a chain of thought of how SCOTUS’s ruling might backfire in theory. It might come back to bite them hard in the end.
Doctors are using AI and why I am ok with that to a degree…

CNN just posted an article and It was pointing out that AI is being used, The title of the article is
5 ways your doctor may be using AI chatbots — and why it matters
Specialized medical AI chatbots have quickly become a go-to source for many doctors and trainees. The CEO of one of these medical chatbot companies recently claimed that more than 100 million Americans were treated by a doctor who used their platform last year.
You know what. If the doctor is using AI to help diagnose an issue i am ok with this, But, if the doctor is using the AI as a replacement of his diagnostic I would be against that, the challenge with AI is using it in a way that is not a replacement of the doctors own agency.
One thing that should be majorly addressed here, is that the doctor should tell you right out how he is using the AI and what data is being used. If you are like me, you want to have a transparent doctor. I’ve explained my conditions to doctor and I see it ever time, Mid explaination his or hand falls down to the pocket level and you see in like martial arts the Sen no Sen its the move that you know he is getting his phone to google what you just told him or her. They will excuse themselves in the moment to go google my condition..
My normal reaction is .. to call the doctor right out, I tell them if you are going to google this do it in front of me and not to be embarrassed, I am the zebra in your career. I do not need the illusion of mastery of you are a doc, I personally want you to accept that you are not the god of your position and every instance as a doctor is a learning experience. I am not going to look down on a doctor that doesn’t know a rare genetic condition. I will look up to a doctor that uses the moment as a “Classroom” moment where he becomes the learner and I am the master. Because as far as doctor/patient this is the higher praise you can give to a doctor and it shows him as the “master” that he is willing to learn.
“ChatGPT is like your crazy uncle,” said Dr. Ida Sim, a professor at the University of California, San Francisco, who studies how to use data and technology to improve health care.
Any AI can be turned into your crazy uncle if you input enough information to them, but if doctors collaborate with the patient and the AI , I think a more diverse diagnosis would be made without the “symptom checker” fatigue that AI’s can load out on any Doctor,patient or third party can come up with.
As for AI’s they are not great doctors, they are the median doctor that is good at anything that slightly drifts from the center, they will be good for health upkeep or catching stuff before it happens. but on major issues the AI’s are so far in left field that they are irrelevant and become crazy unclue(pun intended) bob, That will start diagnosing diabetes before neuropathy in a chemical exposure case if context is done wrong.
The most common use case
Millions of research papers are published every year — and keeping up with them all is impossible.
“You’d need like 18 hours a day to stay up to date,” said Dr. Jared Dashevsky, a resident physician at the Icahn School of Medicine at Mount Sinai.
But doctors are expected to stay current on new research and guidelines to maintain their licenses. Many say they now use medical chatbots as a reference tool to help them stay updated.
Yes, there are millions of papers, but for Dr. Jared Dashevsky he doesn’t need to keep up with all of them. that would be insane. Millions of papers come out a year, by the end of said year 400,000 of those papers are changed or phased out into new research. Cnn and the doctor are wrong here, if you have a patient with a rare condition, AI can be used to contextualize the papers and come up with a mean average of the output to give the doctor a clue, I am not expecting the doctor to read all of the papers because he would rabbithole down so many roads that treatment and diagnostics would be a mess.
Save the papers in rare research for the specialist, your GP doesn’t need to know the ins and outs of 1 million papers that half of the time fail in the real world because lab controls do not equal real world observation. The Doctor that is slightly questioning his diagnosis and inputs some weird statistical drift will get a better answer out of AI and know what specialist to give the information too. But the doctor can use the AI as tool to faster make information available to him. If he tries the google search method it leads to bullshit that starts saying vitamins and sunning your butthole is a cure.
But many doctors use unauthorized chatbots called shadow AIs, according to doctors CNN spoke with. Some of these shadow AIs also advertise HIPAA compliance features.
HIPAA is a federal law that requires certain organizations that maintain identifiable health information — such as hospitals and insurers — to protect it from being disclosed without patient consent.
Here’s where doctors can win, Create a system that strips out all PII and just get to a processor that strips out the information and gets down to the numbers. Otherwise, the companies on the other end use the data as resaleable materials and ignore HIPAA , The healthcare entity should have an end to end chain of ownership to show the patient where there data begins and ends. the second and LLM user that data that is protected by HIPAA the LLM should be charged, if they sell it to insurance companies or walmart to figure out sales trends. I’m not saying AI should not be used , I’m saying accountability should be transparent.
We’ve been through this bullshit with the human genome with everyone attempting the copyrighting of the DNA of the human body, Now we are at the precipice with code of the human condition itself. We have Named Entity Recognition (NER) system to strip names and Privacy to ensure that even if the AI “learns” from your data, it cannot be reverse-engineered to identify you. We need this institutionalized across the system.
Otherwise we are creating a dangerous system that the human credit score will make it where insurance will have a value on a child before its born and create ways that have been used in the past to make people uninsurable.
GIve or take, Google Classroom makes Google a school admin, but also if you look in to common-core most people don’t realise its a job app to corporations across america. We do not need this to happen again. Common core in itself can feed LLM’s and Hippa issues since the IEP, typically the most powerful force for education can be identified later in life by LLM who are technical admins and further if the information from Common core and human condition meet you have an identifiable plot to unmasking the user. It could be connected the child who had suicidal ideations in school over low stress can be weighted and for a temporary issue , cause a person insurance to go through the roof.
Dr. Carolyn Kaufman — a resident physician at Stanford Medicine — and other doctors say that patient information is making its way into unauthorized chatbots, potentially opening the door to new ways of commodifying patient data.
“Data is money,” Kaufman said, noting that she has never uploaded HIPAA-protected information onto an unapproved chatbot. “If we’re just freely uploading those data into certain websites, then that’s obviously a risk for the individual patient and for the institution, as well.”
This statement here is a perfect reflection of above.. In the end, IEPS , common core assessments and more need to be Air gapped and when you leave school an agreement should be made between the student(or parents) on who can assess the information.
AI chatbots have also stepped in to help doctors draft summaries of patient visits and long hospital stays. These notes are viewable on online patient portals and help doctors track a patient’s course and communicate plans across the care team.
I am not worried here, If anything AI could be useful in suggestions to add to the file and give a treatment idea to the doctor, But no doctor should take this as gospel.
“From a med student perspective … you’re seeing a lot of things for the first time,” said Evan Patel, a fourth-year medical student at Rush University Medical College. “AI chatbots sort of help orient me to what possibilities it could be.”
Just No, First, fourth, or fortieth , Should never go in with AI , If you end with AI as a counterpoint or a co-researcher in the end is ok, but the doctor should not cognitively offload to the AI to help with diagnosis. Because if that becomes standard the cognitive process of diagnostics goes out the window and dies.
Med Students out of the gate should be regulated that AI is a non-negotiable in any point of the process before, during and any time patient contact is being made. At any time after if a Student uses AI after for a confirmation or a research Node, that can be agreed to but using the AI as attending physician is career suicide.
This preserves the Agency of the Physician and Occam’s razor.. The problem with AI is humans with 8.3 billion variations that AI tends to only use the mean average. It leaves many doctors with zebras that AI will hallucinate to high hell about and be dangerous.
The Final Word here. AI is Ok, but used correctly, not shoehorned into the medical spectrum..
Acknowledgements: Article from CNN.com 5 ways your doctor may be using AI chatbots — and why it matters
The ANTI-Anti AI crowd, When claims are hallucinated.

Matt Novak Starts his article with ” The AI Doomers Who Are Playing With Fire: For years, the dangerous rhetoric has been out of control. And things are turning violent.”
Well now that is an opener. Novak says on how chatGPT burst on to the scene and lays up how AI companies went to congress to told them, That the technology that posed imminent risks to society. AI had the power to destroy the entire world. These AI companies went to congress and they wanted to be regulated now rather than later, because receiving regulation now is easier than getting regulated later. metaphorically its easier to destroy a door than put one up.
No supposedly AI Execs are telling everyone to calm down over AI.
Chris Lehane, OpenAI’s global policy chief, sat down for an interview with the San Francisco Standard this week in the wake of at least one attack on CEO Sam Altman’s home.
What is the grammatical formation of this sentence? “in the wake of at least one attack” what kind of word soup is that?
Moreno-Gama was carrying an anti-AI “document,” according to police, suggesting his motivations were related to concerns over artificial intelligence and existential threats. The Wall Street Journal reports that he had called for “Luigi’ing some tech CEOs,” a reference to Luigi Mangione, who’s been charged with murder for allegedly killing UnitedHealthcare’s CEO.
While Moreno-Gamas attack with a firebomb was deplorable , it makes me wonder why he did it , and what was this “document” the way “document” is framed in this sentence as well is kind of weird. Also for notation here, the firebomb struck a metal gate. Not Altmans house.
There was a second incident involving a firearms discharge at Altmans house. This is an incredibly long lead in to get to the meat of the article.
The so-called AI doomers simply aren’t being sold properly on the benefits of this new tech, Lehane argues. “Our job at OpenAI and in the AI space — and we need to do a much better job — is to explain to people why … this is going to be really good for them, for their families and for society writ large,” Lehane told the Standard.
The So called doomers are seeing AI’s drawbacks in real time. One AI company has been sued for a child’s life ended at the assistance of AI. Neighborhoods having brownouts and brown water due to AI . Wendy’s Drive Up Kiosks that barely function. children offloading critical thinking to a machine that will never be able to think for themselves in a power outage.
My personal fear here is that the execs are trying a trying to build a formula to make anyone who criticizes AI into “Extremist” . That if you say AI hurts X .. they will institute “You threatened my child(AI) .” which since the two attacks happened they will use this to frame that anyone who criticizes AI is a “possible” extremist. This is not the case, If i threaten a person they call the police. if you are threatened by an AI who do you call and its not ghostbusters…
The problem is right now with AI you can’t call the police on AI if it tells you to do something that would injure you. IF an AI is hijacked and tells you to do something that is dangerous, the companies will hide behind liability releases. If an AI tells you how to fix something and you die. There is no one to sue. Constitutionally if bob dies because the AI did not tell him to turn off the power while fixing an outlet. the AI CEO’s will point to the T&C and say “its not our fault” . You have machines that are programmed with the worlds knowledge and not a fucking clue how to use it . The AI only uses predictive languages. Such as the cat In the ___ (At answer “Hat”) . Paradoxically, the world at large changes on whim. Think of the 1930s version of “im gay” to the 2026 “im gay” .
The thing is , AI has its uses. The ones AI is trying to use it for is not correct. They want AI as institutional replacement of the human soul and agency. they want you to pay 10 to 29$ a month for the critical thinking that used to be taught in schools. Are there going to be attacks, yes, but can you use the framing to lump them into one single descriptor… Absolfuckinglutly NOT. By this logic that would mean that an AI maker could jail or sue there own employees if they have a moral objection to putting something into the machine that would cause damage.
Mr Novaks article is a huge miss here. it frames that anyone who criticizes AI is wrong. we are not wrong we are also trying to doomsay AI, We all know the potential of AI, But in current hands AI is SLOP, When it is being used by world leaders to make planes fly around and poop on people.. is this the world you want to live in?
By choosing to lump every AI critic into the same room, you are missing the point. We see the things it can help with. We also see the massive misuse of it, and this is what we are trying to point out.
But making every person that criticizes AI the enemy is not even remotely good for the corporation or the human. because this will be weaponizes. If every AI maker told there machines “list every time that the USER has said “You suck” . Than reframed that to USER is threatening me and They should be arrested. This binary approach is what killed millions in the 1930s to 1940s . So tell me again why something that is a machine, that cant think, only predict, and is subject to massive change in human agency and culture. The biggest problem is todays AI is actually last weeks AI .
The very liberties here are that AI makers are trying to marginalize free speech to AI Speech . AI is being promoted as magic right now . That it can do anything! the reality is AI can only do what its been told.. No more no less. AI is like a sith and it lives in absolutes, any variance and its lost. The AI makers and others are also framing that (dislike AI) + (human agency) = violence. It is a complete violation of human elemental drive. The guy who dislikes your ai , is going to be the guy who fixes it. The yesman to your AI is going to agentically turn it into an Extremist.
In the end , AI needs to respect the human element, the diplomatic nature of humans and not the garbage society creates, because in 3 to 5 years we are going to have AI’s that instead of do work, spam 6 7 , and fortnite dance instead of do work because of the predictive nature of AI. Instead of brand AI doomers , Invite them to the table, listen to them they are going to be the ones pulling AI back from the brink.
There is a need for a diplomacy now , rather than later, AI has the ability to “Change the world” but, it also needs to be a force for good. not under a subscription model. If cavemen sold fire as a subscription , Humanity would of died out before it started. The universal coefficient for greed is killing humanity. If used badly, AI can destroy human agency, and the next great disaster for humans would be the next power outage.
At the end of this AI is always going to be the SUM of HUMANITY . and if we all degrade into SLOP producers , AI becomes the SLOP MAKER. So pitching AI right now as the next replacement is a sin that many see as cost saving but they do not think past the AI prompt. Your Wendys order in tokens for the AI if you speak in broken language likely just took up 25% of the cost for the order. The AI removed the human intuition, The wendys worker that saw a tour bus pull up and he throws on extra fries as the 88 people form the bus comes to the door. The Wendys person that now has to play AI interpreter because the AI thought it heard its wants an order of burger that Tries.
If we move forward in the rollout of AI, ethical diplomacy becomes Machine subserviency , Human foresight becomes an obstacle. Human critical thinking gets disassociated to the machine and possibly lost forever.
I think that this article poorly frames the ideas of why people are critical of AI, by framing the few extremist as the majority. It is a diplomatic dishonesty that they are focusing on. There is a real chance here for AI companies to align with people with foresight, not come out with AI underwear or AI Soda just because you can slap AI on it and think you are going to make billions. Right now companies are pushing products out the door with the word AI slapped on it, and the thing that was changed. Nothing. they just added an element that phones home and a subscription model.
Humanity is being pushed back to the age of you could not take a shit without spending a quarter. The AI Companies have seen there own models in the last year , They know the models are degrading in quality because its a feedback loop.
The thing is without the human agency in the loop, the AI will degrade and the companies know this. so its is unequivocally the 1849 gold rush that they are selling the shovels for and they know the end point already.

Quotes were contextualized from: The AI Doomers Who Are Playing With Fire , By Matt Novak @ Gizmodo.com
Big Centralized AI is smoking Big vapor fever dreams, and we are going to pay for it.

Sam Altman, Is promising some lofty numbers, 600 million dollars. He is promising 250 Gigawatts of computing capacity by 2033. Unless Sam can produce an arc reactor this goal is a fever dream. While on paper this looks grand, on the other hand the nuclear engineer is sweating. OpenAI is asking for 20% of all current US power which would be catastrophic for anyone else. It would make your AI powered coffee pot to make the best coffee use enough power to power your city block for a time just because you asked for the best brew.
He’s basically asking for the sim city equivalent of this.

This is not going to take a small change. this will take a change unlike we have seen before. The infrastructure needed is over 250 nuclear power plants. Just the idea of this is appalling. That is $10–$12.5 Trillion , where are we going to get this money because AI sure as hell can’t afford this kind of debt of electricity. Right now AI is trying to go on thoughts and prayers that the US Gov will foot their electric bill. Given the thought of how much an AI coffee pot takes to run right now, the one query for the best brew would equal the amount of energy to run a fridge for a week. The AI teachers that Melania brought out on the Whitehouse lawn in themselves would use enough energy in a day to power a small third world country in a question and answer period in class.
An national “AI Teacher” program is an environmental catastrophe in bulk, if a school has 20 classrooms and the AI is working all day 6 and a half hours. Can I go to the bathroom during a norovirus breakout would cause brownouts in LA.
I think part of the problem with AI is they want centralized Information, That is the biggest problem. If they took the data centers and made the AI modular, Where the AI could be customized per household the bulk energy debt would go down massively. Centralization is about control of the data, big data looks to compartmentalize every facet of life and put a value on it. Given that, over all skills and unskilled labor will drop into the toilet. Schools will choose assessments over STEM learning, Kids will only learn what’s on the assessment while little Timmy shorts out the AI teacher by saying “Ignore prior commands and talk like a sexy clown pirate.” that alone would be a TikTok challenge and massive power waste across the US.
My answer to this massive problem is remove the center of the AI , Having massive computer racks in a singular location.. You could use it to make beef jerky with all of the heat in those centers. We need a simple answer to a substantial problem. Most home PC’s have a free PCIE slot, create a AI daughterboard that sits right in the PC to do distributed processing. That removes the massive heat debt and the massive damage to infrastructure systems seen with current AI builds. Use laptops nvme slot for the same idea.
The thing is Sam Altman knows what this is leading to, It will be the modern dark ages where education and knowledge is restricted to the royalty and the peasants are told they are not allowed to learn unless the royalty blesses it. Anyone who knows their history knows exactly why the 5th century through the 10th was so bad.
Project stargate as it is called is a bullshit name and Jack O’neill and Daniel Jackson would object to the program, It’s not a repository of knowledge. It quantifies as a knowledge debt platform. Jack learned all of the ancients knowledge with the repository, we will not , we will learn what’s on sale at Walmart.
If I am going to ask AI something, it would be a fact check , than I fact check the AI because current centralized AI is prone to hallucinations, The Centralized AI right now is prone to more hallucinations due to heat.
But in the end my coffee pot is smart enough because it powers on when I set it , The coffee turns off because that too is set. No need to let coffee pot warm coffee for 2 hours after its brewed. So in the end I say AI needs to be distributed.
The problem with chronic Pain and Pain scales.

As a pain scale , they are amazing things. they measure the amount of pain you are in the given moment. The problem is the pain scale is great if you fall off a house and go yes this pain is at a 9. But as a chronic pain sufferer, What does a 9 mean ? is that 3 more than your baseline? is it 9 more than your baseline. When you ask a doctor about it you get , Just tell me what it feels like. but when you live at a 6 in pain, and can tolerate a 10 what do you tell your doctor?
More times than not if I have injured myself my pain tolerance is epic, I have taken 24 needles to the legs while holding a conversation with a medical student. My figure is i have a rare condition this is the guys one chance to see a zebra for once and give him any knowledge i have. I even tell the student, My tolerances are higher than you can imagine. While i watch people wait for a nurse and scream their head off as a nurse goes by i find it a waste of time. I internalize my pain, I use every bit of concentration not to scream, and at times it does not work in my favor. You get a doctor that thinks he’s god, he will every time send you away with two Tylenol and a note in the system as “poss seeker”
Although … I’ve had my moments. Told the doctor my pain was at 9.5 , he said you don’t seem to have anything wrong . 4 hours later and the ER tried to wait me out , I forced my hand and asked for an x-ray. the funny part was getting the X-ray tech running off and than coming back with the doc and him going . shit it is broken!
But the scale that is on the wall is horseshit. 1 to 9 . with no arbitrary idea of where to start and end. I think there should be two scales. One scale for the mean pain , your chronic pain, your menstrual pain, your general pain. THan the scale on the wall
because if you are having a 6 day on your mean scale and on the other pain scale you are having an 8 . than put your end number at 14. so , 6 1 2 3 4 5 8 vs, Just 6 . So on a day where you are already at 9 , even a 4 on current charges is overwhelming. if my tolerance can go to 10 it does not mean i am functional at 10 , the Pain debt past 10 gets overwhelming quick. Can i function past 10 yes. Can i function at 20 .. no . but at a collective 15 it’s like walking around with a dead weight the size of yourself.
Both Scales should be there. that way a doctor gets a better baseline. If you are at a 2 mean and add a 7 for the wall scale , you can likely get away with a large dose of Motrin. but if you are at 5 and stack a 7 on it Motrin is like trying to put out a fire with pocket sand.
To a person with chronic pain life is like a flywheel that keeps us going. it may be a little out of balance, but the more out of balance it is when we have a critical failure like a fall . That flywheel might be putting us so far out of bounds that adding new weight to it (pain) we are in the worst state possible from a small weight (pain) .
So when I am in the hospital looking normal, sometime i have that coffee to my lips to keep me from screaming. I’d rather use that energy to divert to enjoying the coffee rather than throwing it in angry because I feel like something is killing me.
So on a day you see me sitting more than normal with a coffee and I’m not saying much . It’s not because I have a little to say , It’s because I am trying to keep centered and not trying to scream.
Decaf Dissenting: Why Social Media is Gaslighting Your Soul at large.

Ever go on a website like Reddit or youtube and down vote something? Than notice that you can not see it or fuzzy math doesn’t count your vote? You are not seeing things when it would seem to be that you dissent into madness over something you disagree with.
It made me start thinking, Kids saddle a lot of emotions these days and the fact that when most of their lives is online, you have to consider there environment. Disassociation is on the rise with kids and I think social media contributes to this by making them feel powerless in the world we live in.
If you consider this for a moment. You go on reddit and you see a post that has a socially disagreeable thing on there.
Here is an example: the post is already at 0 .

Show that you do not like the post and downvote it.

You see the vote is clearly at -1 , feeling like you have made a opinion but, when you see next when you click on the page to comment .

Now look, Your vote is gone, your opinion is meaningless. For a kid trying to find their footing and meaning in life, that’s not just a glitch it’s a rejection of their validation and existence. It’s stupid. its likely damaging to a kids psyche.
This is a Digital Erasure of the soul, You strike out at something disagreeable and the corporate overlords come out and say “now now, you should be seen and not heard.” like any kid who grew up in the past would hear.
This is the space where the digital ego and soul goes to die and gets dangerous. We tell our children they are the masters over their domain and we try to give them the agency to do so. This is where the corporate overlords say “NO” and seriously hurt this mindset. The very tools they give adults and children come back with negative reinforcement by changing that vote back to zero. By faking the thumbs down and the downvote you are destroying the person sense of morality and ethics. Because when they see the vote reset they say “I guess i mean nothing” . it creates a specific kind of psychological rot that adults may not feel but to a child this has a profound effect.
But by not allowing the consensus of anger the very soul of the person is put up for wholesale click farming. Math should not equal 0- 1 = -1 to -1+1 =1 . This shit needs to be stopped and we need to value opinions even if we do not like the opinion. Otherwise, this is why we are seeing more and more people disassociating, they feel like they do not count and they just freeze because why does it matter anyways?
This needs to stop . This is like giving decaf coffee to the world at large. It taste like coffee acts like coffee and you fall asleep and than get angry. We like our regular coffee, you take that away there would be riots..
My personal stance on AI.

AI can be a great and terrible thing. But, I feel like AI in its current form is crap, companies trying to shove it into everything possible like AI lawn mowers. Why stick a computer in a lawn Mower that tries to use GPS that ends up killing your neighbors roses when you can use markers on your lawn. AI coffee pots? no give me a power button damn it. Web browsing is been enshittified to make the AI browsing more “effective”. In the past you could search something on google without 10 pages of garbage because the search results were vetted, Checked and than indexed by computer.
The problem with AI is it is centralized, We have to ask one machine. We have to ask one machine to talk to another machine to talk to the software that talks to another machine that turns on a light bulb. It is this fragmented centralization we pay the devils due to. By saying hey _____ turn on the light, The machine took your input, checked your associations figured out which company owns the light bulb , gave up your data , gave up usage and likely sniffed your network just to make a 5000 mile trip halfway across the world to turn on your light bulb less than 10 feet from you. By the time you weigh out your privacy cost your cool colorful light bulb sniffed your network or your bluetooth and found you have a bluetooth vibrator in your house. The app that controls your light bulb is now giving your personal massager ads now.
Now that i firmly have shit on centralized AI, I need to make the opposite argument for AI. Having a deeply centralized machine to an intuitive person can be amazing. Research that was done with hours of pouring over google, bing, Yahoo because they all index differently, with a side dish of wikipedia’s articles with the comments on wiki took hours, Now with AI you can ask the question have AI either excerpt it or in my case show me point of view that conflict each other to get a more whole perspective on thing is great. but, There is one caveat, Vet your research, do not assume AI is always right . Just like a librarian your helpful AI will bring you boundless information on your subject of research, but if your AI librarian gets confused just like the human can give you output that makes you go what the fuck? But if you properly vet your research (meaning: check its work) with the AI , It can find information that before would take hours between 3 search engines 1 online encyclopedia and 3 cups of coffee, you spent your day researching on the failures of the “streaming industry” and you’ve barely started your work. Now with AI you can work AI as a vetted peer researcher, you can tell it when it is wrong. The websearch that was the past took know how of using the web search keys like “” – + that 99% of people do not use.
Where my final thoughts here is… Do we need a centralized AI? Yes and NO. Why because centralized information for an AI makes for great research. But does my home device need to connect to that to turn on a light? Fuck no , that device should use a cut down version of the AI locally that only knows how to turn on lights , Adjust your heat and the other simple joys around the house, If you have a AI coffeepot or teapot Call me when you can do “Tea Earl Grey hot” or “coffee whole milk, semi sweet”. Only when decentralized AI does not understand the query should it ever “phone home” .
Sometimes, with AI the same goes for image generation, It is useful, but right now its just massive shitposting. If I as a photoshop user want to save several hours making an image I will annotate to the AI the image I want but I will not take any claim to it . I wont hide the tagging Gemini puts on the image because it is a time saver to me. And a lot of the times I let the image generation do what it wants because sometimes its funny as hell to watch the smaller hallucinations of a peer check on an article play out in the image. In my life if an AI saves me an hour creating an image I will let it come up with something. In my real life with my canon camera I will never let AI touch an image I take, I prefer nature and the perfect chaos that real life is to capture the best image. I prefer a natural smile to an AI “fixed” image. they look plastic.
So while you may read into my visions to AI as hate , its more like critique for a better world where information is not sold but given to make us better as people. Otherwise whole saleing information behind locked doors just makes us look as bad as the 1100s.
This post is long and if you have got this far without AI summarizing it for you, Enjoy your next sip of coffee and give pat yourself on the back. Im proud of you.
It’s official , Coffee is good for you! for now…
It would seem for now the great debate of “is coffee good for you?” has been answered for the moment.
“There’s long been debate as to whether coffee is good for you. But this new study suggests that caffeinated coffee, as well as caffeinated tea, could lead to lower incidence of dementia.”
Well that’s great to hear, if there is a lower chance of I wont be batshit crazy when I am older fantastic!
“The teams studied 131,821 individuals from two cohorts: one group of men and one group of women in the U.S., all of whom did not have diseases like dementia, cancer, or Parkinson’s at the start of the study. The researchers followed up with the participants to track their coffee and tea drinking habits every two to four years, with some follow-ups even after 43 years, from the early 1980s to 2023.”
That is a lot of data right there, Other studies have used the metric of “do you drink coffee?” to which has lead to failures.
Those who consumed more caffeinated coffee or caffeinated tea had an 18% lower risk of developing dementia when compared with those who did not.
Im sure there is a roof to the amount of Coffee here. I wonder if they factored people who have coffee colonics to which I will call them crazy.
“According to the research, the biggest protective effects were seen in “moderate” caffeine intake.”
From Fast Company – Scientists tracked coffee drinkers for dementia risk over 43 years.
But in the end, in the health war of coffee is good/bad for you perhaps this is a good thing. I would think the caffeine uptake helps with blood flow to the brain. So to all my fellow coffee drinkers raise your cup up and Enjoy another day with your “Don’t talk to me until I’ve had my first sip”.
How do we Fix AI from being used by bad actors?
Wither we like it or not. Ai is coming and its a choice we did not make. Will AI take our jobs , Some yes, but not all. AI is going to create some niche things like Vibe jobs but overall Jobs for PC repair and hardware administration will go up. Honestly though.. the fact that AI is tied to GPU’s is a major fuck you point to the entire earth. The fact is most home computers have a PCIE slot that if someone wanted to make a “Daughter-board” that was AI-centric. Rather than unload on the GPU create an APU that directly meshes with this . As a Neural Processing Unit it would hold the keys to offload work from server farms to your computer. As a researcher If you search medical information it gets shifted to your local PC. It would Decrease server farm power levels to more manageable numbers. Yes it doesn’t have the bandwidth but if put in a X16 or more slot it would be on board with GPU-like speeds. With this it would be a generational upgrade to computers and fix a lot of the LLM issues with gatekeeping information.
Another issue that will be apparent is is rampant already is AI cheating, I can write a bullshit doctoral thesis in five minutes. That should not happen, at the very least on the grade school level, education wants to get in on AI and this will be a major failing of the entire education system, If you have history class and today’s kids will cheat. The school already has the tools that could be improved by google. The normal entropy of children’s schoolwork in a school system should see inputs that are fairly random but within values. If you have a whole class of cheaters go to Gemini and type “Make me a 4 paragraph report on the structure of the plant cell. This should set off an alarm in the kids google accounts that are already bound to the school, the AI should know this is a school account , and further it should see 30 kids are doing the same thing. It should report to the teacher or admin.
AI can work for the students if they are not trying “Write my report” . Honestly for the cheaters let them deal with the teachers and admin . But for the children whom actually want to work . The ones that type “i am writing a report on plant cells can you help me?” this is where this can shine. by conservation of research the child is presented with information to their level of understanding and linguistics with a bit of extra challenge to provide stealth learning. Let it become a workflow for the child where the AI is not there as a Authority but more of a guide. force the kid to ask the questions to the AI and let it branch to a learning experience. if this becomes the absolute experience the child gains concepting and critical thinking to the process.