Skip to main content

Opening a new chapter at gid.space

Welp, It's been a while since I've made a post. I've been busy over the last year, what with working at NTH, Inc, learning computer sci/programming, and reading a ton of great Sci-Fi.

At the moment, I'm working on a web site: gid.space

Yeah, it's a weird url, but it's short and I like the idea of it standing for a virtual place I can put all of my crazy ideas, and tie everything together.

I may not continue to write on Blogger in the future, since I'll be trying to do short posts on the homepage of gid.space

For now, thanks for browsing through my strange ideas, and look for me at my space. The gid.space *puts on way-fares and walks off into the sunset*.

Comments

Popular posts from this blog

Tech Archaeology: Unearthing the Artifacts of a False Prediction

Greetings. This is going to be a shorter rant. New year, new me! Anyway, I was inspired to write this after I caught myself falling into a usual habit: investigating the validity of a prediction which claims that a technology (it could be anything) will take over in the future. I'll start from the beginning. It all started when I was dutifully studying for my Databases class. While reading the textbook, Database Processing  (13th edition) by Kroenke and Auer, I came across a passage that was summarizing the history of database processing. Being that this book first came out around 1977, it has probably witnessed very few shifts  in the popularity of database technology over its existence; namely, the rise of Relational Model and its subsequent dominance. Never-the-less, in a table that describes the emergence of database technology, there is a row for the "XML and Web Services" era (after "Open-Source DBMS" and right before the "Big Data and NoSQL"...

Parallelism and Task-Decomposition: An Introduciton

Introduction Since I’m on holiday break from University, I’ve had time to begin investigating parallel processing. I’m going to try and share a little of what I’ve learned about the technology and how programming languages can leverage multi-core CPUs and GPUs. I’ll finish up and explain a bit about task-decomposition , which is an important aspect of writing algorithms intended for parallel processing. My investigation began one night as I was cooking dinner and watching a talk by Rich Hickey, the inventor of Clojure, titled “Are we there yet?” . At one point, while Mr Hickey was discussing parallel processes, the question of “how does a GPU process pixels?”popped into my head. Surely that must be parallel, right? Since my “Intro t...

One AI to Rule Them All

In my last post, I mentioned that I was interested in looking across the field of Computer Science to see if there my be other areas worth working in in the future, besides web-development. Primarily, I'm interested in a more powerful run-time environment. So Python or Clojure seem very enticing. A safe first step would be to learn about the architecture of server-side code, to deepen my understanding of Web-stacks, and also to satisfy my curiosity of what else is out there in terms of languages, VM's, and code paradigms. Another motivation to look outside of web-dev has come more recently, in the form of some sage advice from an interview with Elon Musk . In it he states that there is too much talent working in the internet space [  55:00 ]. So if web-dev is overcrowded, what would be a worthy Comp-Sci field to pursue? In the same interview, Musk also alludes to the seriousness of the approaching revolution taking place in the field of Artificial Intelligence. He desc...