at one job, they complained that I gave TOO MUCH INFORMATION in
my bug reports. Too much?* TOO MUCH?!* How can you solve a bug
if you're missing info?* Gotta know context. Sometimes the bug
isn't where you think it is.* After all... if it was an OBVIOUS
clear and easy bug... somebody probably would have spotted it
already. It's fun to be the first though. The best cause for
bugs?* Video memory.* Bad video memory.* Nobody ever looks
there.** I've seen that problem from the days of Hercules
monochrome amber screens all the way through to
super-deluxe-duper-3D-rendering-the-Universe-in-a-box video
cards. One of my favorites?* Laptop didn't work right anymore
because the SOLDER CONNECTIONS to the video chip MELTED from
games utilizing the video memory for matrix transforms... and
the solder didn't see THAT level of heat coming... melty
solder... loose connection... laptop? No workie no more. Sorry -
the GPU.. the GPU.. not the video memory... gpu got hot.* The
gpu utilizes the video memory for higher level operations
because the gpu is capable of a type of multidimensional
mathematics that regular CPU+math coprocessors can't.* Turning
2D into 3D is hard to do when all you have is the equivilent of
a piece of paper and you're trying to do origami with a
toothpick and a straw as your only utensils.* [toothpick =
electron, straw = electron hole] but I don't wanna get into my
complaints about the flat layout of memory.* Drives me nuts.*
Memory needs to be in a 3D box to allow for cross-connections
easily along with a little analog uncertainty built into the
logic systems... but we'll get there.* we're still building
computers based on old models from 50 years ago anyway.