I needed to install WAMP on my PC, to support some Amazon developments; I needed a screen interface to my database. Here are some notes and links. I talk about installation, and port conflicts. I might expand the install notes to include advice on mysql security. Read More
Flume is a distributed, reliable, and available service for efficiently collecting, aggregating, and moving large amounts of log data. It has a simple and flexible architecture based on streaming data flows. It is robust and fault tolerant with tunable reliability mechanisms and many failover and recovery mechanisms. It uses a simple extensible data model that allows for online analytic application. I need to check this out. Read More
Here are some notes on using apache. This talks, a bit about Access Control and redirection. I need to test if any of this content was duplicated elsewhere in the site. Read More
The Apache Hadoop project develops open-source software for reliable, scalable, distributed computing. I became interested in this when looking at true 24/7, follow the sun computing. I published a blog article after reviewing Cloudsoft’s Monterey middleware. I thought it was pretty neat although I haven’t used it in anger. Read More
This should be easier, but I think made harder because I corrupted my Firefox applications helper list, and then chose the wrong way out
My problem, the browser tries to download index.phtml, in retrospect because, I believe that the module wasn’t properly installed into apache. Using Ubuntu 9.04
The Right Way
apt-get install apache2 php5 php5-mysql libapache2-mod-php5 php5-mysql mysql-server
I am running in a virtual box so I needed to create a /etc/apache2/conf.d/fqdn. There wasn’t one on my system, it needs the line
I suspect that fixed it, and that my errors were caused by not initially correctly installing libapache2-mod-php5. I have proved that my install works by using Opera. Now to remove the .phtml line from the firefox browser and make the vm a server.
Other’s I’ve found and used include
A number of people publish top 10 lists, you can use google to find them.
I am interested in writing or implementing a url-shortener, for use inside the firewall, or implementing through an API. It seems quite simple until you consider performance at scale. My first thought was to use a database with an identifier/sequence data type as the key to the hashing algorithm, since its best to keep a record of the URL’s issued so if people ask twice for the same url it can be reused.
The second part is resolving the URL on request. Is this just an apache solution with a very large redirect file?
I used google: url shortner howto, to see what it came up with.
Two interesting finds included
- http://blog.cullect.com‘s article on how to build a url shortener, but actually its about using their SaaS
Idea No 1 was to use a database identifier type and hash it, I was going to investigate if we could use a web server redirection file, or if I’d have to write something using a db retrieval. However both these the database transactions introduce serial bottlenecks and inhibit scalability.
So can pre-allocate the sh.urls in blocks , and allocate them using hadoop? Is hadoop a sensible solution for “find me next unused”, or perhaps it doesn’t matter; its just find me one unused slot?