instruction stringlengths 4 105 | output stringlengths 8 56.7k |
|---|---|
Hey hackers how many hours do you sleep? | trickjarrett: As many as I can fit so long as I'm up by 6am. Usually 6-7ish. |
Hey hackers how many hours do you sleep? | mikeyur: Depends on the day, usually 6-8 and sometimes a 1-2 hr nap that day as well. |
How do I handle a million hits a minute? | davidw: That's a lot of traffic. To handle that dynamically would probably cost you a pretty penny, even with really fast code (Apache/nginx/whatever modules in C, for instance). What are you trying to do, exactly?My laptop can pump out a static page at about 4000 requests per second. A million requests per minute is on the order of 16,000 requests per second, so you'd probably need multiple servers running really fast code. Since you don't explain what you want to do with the data, it's hard to really say anything meaningful about how you want to handle it dynamically.Do you really need to start off handling that much traffic? |
Hey hackers how many hours do you sleep? | entrez: 5-8. 2+ hours of exercise a day enables better concentration for longer periods imho. |
How do I handle a million hits a minute? | ErrantX: I've been looking at something similar (not on that scale admittedly) recently.Either I would go with a generic cloud service to host on (which should scale to handle your hits) such as EC2. Butthe cost could be astronomical.Or you could go with something like 3tera & the custom cloud solutions and make your own server system: I'd say 4 or 5 servers behind a load balancer.However w/o knowing more about how data will be stored I cant advise on how to keep the database under control. A million database writes per min is not going to be pretty - even a million reads per min is going to stress it... give us some info on your DB architecture and the process / requirements of those hits and someone will be able to advise more I am sure :)
(Edit: as a hint your probably after somethign like slave db's and memcached - have a look at how Wikipedia handles their data) |
Review our startup - BeatMyPrice.com | bengtan: Uh, presuming the technology works and your data is good quality, how do you monetise this? |
Hey hackers how many hours do you sleep? | ideamonk: 4-6 |
How do I handle a million hits a minute? | mseebach: The biggest issue, from what I can read, will be how well the data is partitioned. If every single write will affect every single subsequent read in a non-trivial manner, it's a really hard problem, and you're looking at really expensive hardware.If you're able to partition the data, it's "just" a question of throwing enough servers at the problem. |
How do I handle a million hits a minute? | jwilliams: Theoretically possible I guess. The current TPCC holder is 6million transactions per minute (mixed workload). But, that was on some, er, serious hardware: http://www.tpc.org/tpcc/results/tpcc_result_detail.asp?id=10...I've personally seen 1000+ tps systems (60,000 tpm), but these were all big iron environments... I've seen commodity based (e.g. x86/Linux) environments up to about half that.However - all of that is just guesswork - A lot of how your would go about this would depend on your data. Are there natural partitions in the data? (time, user, geography, category?).And even if your page is tiny - say 10kb - That's, er - something like 1Gbps sustained (my math might be a bit off)? That's a lot of bandwidth. |
How do I handle a million hits a minute? | newsit: As the other guys said it all depends from your situation and what you are trying to do.
One thing to maybe look at is Erlang and related technologies Mnesia, Yaws etc.
This is a comparison of Apache vs. Yaws under heavy load:
http://www.sics.se/~joe/apachevsyaws.html |
How do I handle a million hits a minute? | swombat: Without much knowledge of the specifics it's hard to come up with a plausible solution, but here are some ideas:* Push off as much of the processing as possible onto the client. Ideally, the browser has a javascript library that deals with all the rendering/updating/etc, and just passes stuff to the server as short commands for which it receives an affirmative or negative response, rather than rendering an entire page for each requrest. This might drastically reduce your cpu usage, depending on what else you do in that request.* How live does the data have to be? Can it be a few seconds stale? If so, and if the data is similar for each user, cache the hell out of it as a bunch of static files served by nginx and the like, rather than dynamically serving any page at all.* Is it possible to do any sharding so that each bunch of users gets its own db server? It's easier to serve 10x 100k hits a minute than 1x 1m hits a minute.Without more specifics, it's hard to provide more specific advice though. |
How do I handle a million hits a minute? | gaius: How much work can you do in the other 23 hours? Can you be asynchronously writing from an in-memory cache to a real database during that time? Or pre-computing anything? During that hour, how much (if any) data can you afford to lose in the event of a problem? How much data per transaction as we talking about? Do you need a relational database to generate your pages or could you live with a linked-list of structs (or a tree or whatever) until you can persist it off somewhere?(FWIW I work on an RDBMS-based OLTP system that handles thousands of commits/sec without breaking a sweat, and a lot more at peak times) |
How do I handle a million hits a minute? | andr: I'd try to build a microsharding system. Depending on the amount of data you want to store, go for 256 or 65k virtual shards and allocate a few shards on each DB server. A hash of the key will give you its shard and it's easy to keep a shard map in memory.I'd advise going with BDB or whatever key-value store is fastest those days instead of MySQL. And an application-server layer to take care of the microsharding.Also, try to put most of the work on application servers (things like joins, sorting, calculations etc), because those are much easier to scale than database nodes.If possible, use in-memory logs instead of writing to the disk during the 1 hour and write the log to disk after the rush hour is over. Consider using a RAM-only database + UPSes, if your data is not that much. |
How do I handle a million hits a minute? | gsiener: Let me guess, you'll be capturing data about where people are browsing in real time? |
How do I handle a million hits a minute? | owkaye: Wow, I was writing a reply to davidw's first post when lots of new posts arrived in this thread. Thanks to all of you! I haven't read any of your new posts yet but I will as soon as I submit this one:The concept is a live interactive system with 100,000+ members online, logged in, and actively posting for about an hour at a pre-determined time of the day (or night). Each person may post 2-5 times a minute and this can add up to 1/2 million hits a minute, but I said I need a million hit per minute capacity because it is possible that 200,000 people might be participating in some of these events. This is a high-profile venture and "bogging down" would be an undesirable occurrence in the middle of an event.Each post from a logged in member will contain a value which typically consists of no more than 5-10 characters. A cookie of 5 characters will accompany the posted data to uniquely identify the visitor. When the server receives this data it will append a new record to the database with the post value in one field, the login cookie in another, a date/time stamp in the third field, and an incremented value in the fourth field to identify the order in which each post is received. I think that's all the data that needs to be appended to the db in each new record.(Note that this is my current concept of the database schema, if you folks know a better way to do it please tell me, thanks.)After the new record is appended, a search is performed and the post field is sorted, then the relative position of the visitor's new database record is identified in the results (example 49,501 records from the top). This 'position' information is used by the script that dynamically generates the HTML, which is then passed off to one of the HTTP servers so it can return the page to the visitor's browser.Since the returned data must be created dynamically based on the most recent database update, it appears that the database itself may be the limiting factor here. I can equip the database server with SSD drives to accelerate its performance but this still may not be fast enough. Plus there may be an issue with database size itself since it will grow extremely rapidly with 16,000 new records appended every second for an entire hour ...I have not calculated the size of such a database after an hour of such appends, and I don't know whether existing SSD drives have the capacity needed to store the volume of data appended at this rate, or even if today's database systems can use multiple SSD drives if more capacity is needed. These are issues I may have to deal with if something else does not come up first that makes the whole concept look absolutely impossible.And yes, I am absolutely going to need this capacity from the start. This is not a system I can "grow into", it must be capable of this performance from the very beginning. Assuming that such a system can be created its cost may not pose a problem since the concept will be very attractive to investors based on the huge revenues it can potentially generate. If there is no practical solution at this time I'll have to re-think the entire concept and try to come up with an alternative that works within today's server / database / page rendering limitations. I'm just hoping there's an existing solution so I don't have to change the concept because as-is it is unique and has unbelievably huge upside potential. |
How do I handle a million hits a minute? | charlesju: If your web application idea is super easy, maybe you should just throw it onto Google App Engine or something and let them handle it. |
SAAS distribution network & face-to-face support. | rgrieselhuber: 1. I would definitely recommend spending as much time getting to know your distributors as possible. Ideally, they will come through references of people that you actually know and trust. Once you have identified some candidates, get something on paper so there are no disagreements, make sure you specify jurisdiction to settle disputes, etc. Best case scenario: the relationship is a success and you will have to evolve your agreement to further suit your mutual needs.Figure out how they (and you) will get paid. Will they contract directly with the client for their share and earn commissions from your firm as well? Or is everything sub-contracted through you? Questions to answer. Make sure you put in indemnifications / consider requiring liability insurance, etc.Having an attorney review at least your first few agreements is a good idea.2. I don't think you need offices around the world. The individuals / organizations you work with will have their own offices / equipment, etc. You're role is to provide a partnership, clear revenue sources and, in some cases, leads that they can fulfill. Systems integrators are usually a good bet for this type of arrangement, even if it's a hosted application. What vertical are you in? In marketing and advertising, for example, you may be able to find agency parters to work with as well.Good luck. I'm sure this is a growing trend. |
How do I handle a million hits a minute? | SingAlong: Process Queing would be the best way to solve your problem. If you are processing a million hits a minute then you might need more than one comp/server. So if you are using databases consider using Facebook's Cassandra project(which is a distributed storage system) for your database needs. This would suit you since it has process queing too. |
How do I handle a million hits a minute? | owkaye: Here are my answers / comments to some of the recent posts (more replies to come). Thanks again to everyone for your help and please keep your suggestions coming:to swombat:
The clients will have javascript so short commands rather than entire pages can be passed. Thanks for this suggestion, implementing this alone may totally eliminate my web server issues.to jwilliams:
You said even if my page is tiny that's still 1Gbps sustained, so by implementing swombat's javascript solution I can (hopefully) pass only data, thereby avoiding the potential bandwidth issue.to swombat:
The data cannot be stale, the response MUST include the results of the database search that is performed immediately after the visitor's last post is appended to the database.to andr and swombat:
I don't know what 'sharding' or 'microsharding' is but if the additional details I've recently posted is any help, maybe one or both of you can tell me if you think it is something that's possible, practical and/or appropriate to my needs?to mseebach:
Yes, every single write will affect every single subsequent read in a non-trivial manner. You said the hardware will cost a lot and I'm sure that's true, but I expect the money will become available if the system I've conceived can be built to function as I'm trying to describe it here.to EVERYONE:Given the information you all have provided prior to the time I wrote this post, I have further refined my approach to this problem as follows. Please tell me if this is the best approach in your opinions:1- I will use JavaScript to eliminate page loads and pass only the required 5-10 characters of data from the visitor to the server each time. This should reduce or eliminate the need for multiple web servers and dramatically reduce bandwidth requirements.2- I will find someone who knows C to write an app that receives the data from each browser, pass it to the database with an append command, pass another command to search/sort the database table, retrieve the results, and pass them back to the JavaScript in the visitor's browser.If this appears to be the best approach so far, I still need to solve the database issue. If you have any new comments or suggestions that might help me find the database hardware or software I need, please post them here, thanks. |
Hey hackers how many hours do you sleep? | hapless: An actual poll from a few weeks ago:
http://news.ycombinator.com/item?id=328901 |
How do I handle a million hits a minute? | khangtoh: After reading through all the other replies, I'm really interested in seeing suggestions on the search and sorting operation that needs to be done with every write.Realizing that if the query returns a huge number of records, sorting and feteching those records will be THE bottleneck. |
How do Flickr, Youtube and other high traffic websites handle file uploads? | stillmotion: I'm not entirely sure about them, but I use S3 and EC2 with SQS so that the user can upload the file, it will be encoded while waiting in a queue, then place itself back into storage. That way nothing ever touches my production server. |
How do Flickr, Youtube and other high traffic websites handle file uploads? | bprater: Sniff around the Amazon S3/EC2 documentation and you'll find pipelines demonstrating what you want to do, such as:http://developer.amazonwebservices.com/connect/entry.jspa?ex... |
How do Flickr, Youtube and other high traffic websites handle file uploads? | mdasen: So, what you want to do is stream the file to disk. The problem that most people will face is that someone is uploading a 100MB video file and your code is trying to hold it in memory. Bad! Get it on disk, then deal with it by opening the file.In terms of a daemon, you don't need one. PHP and other languages can execute other processes. So, you want to convert that AVI to Flash? Save it to disk, then execute another process to convert it. |
How do I handle a million hits a minute? | owkaye: Retric said if I provide data that's accurate +/- .2 seconds that's close enough, but he's wrong (sorry Retric, it's not your fault). The fact is, I must provide each member with the relative position of his post each time he makes one, and the only way I know to do this is to append the new record first, and then do a search and sort to determine where his new post appears in the sorted list.Retric, you said I'm not going to be able to sort things that fast for every new post. Do you know this to be true? Is this an absolute fact or are you theorizing?I'm not suggesting that you wrong when you say this, but I certainly do not want to assume you're right just because you say so -- because unless there's another way to determine the position of the new post in a sorted list I may have no other choice but to abandon the existing approach, re-work the entire business plan, and try something else from a completely different angle. I prefer to NOT do this when the current concept is the best I've come up with.What abut other people's experiences? Do any of you know about data systems that can do an append, followed by a search and sort on one field, at the rate of 16,000 times a second? Or is there another way to determine the position of a record in a sorted list without sorting the db to rceate that sorted list?How about if the database is kept entirely in RAM? I know this will speed things up a lot, but will sorts be fast enough for 16,000 a second without bogigng down?mseebach, you asked "What exactly is the output?" and I'm not sure if your question was directed to me but if it was, and if you're asking about the data that needs to be sent from the server back to the browser, then I think I might need to send only "one character" back to the Javascript. Previously I was going to send an entire HTML page but it seems the Javascript solution could be orders of magnitude more efficient. By using the Javascript approach I can send only 10 chars to the server and receive only one char back from the server. The Javascript can then convert that one char to the proper "relative position" value for display in the browser -- or so I hope.bd, you said I might do better to build my own custom in-memory data structure rather than use a classical ready-made database solution. Maybe you're right. It would certainly reduce the costs over a super-expensive high-speed database solution, wouldn't it? I've never done such a thing before but maybe I can find someone who has. Any suggestions?bd, you also said I do not want to sort the whole dataset after each new piece of data comes in, and you're right about that too. What I really need is to insert each piece of data in the proper pre-sorted position as it comes in. Then I never have to sort the data later because it's being sorted before or as it enters the db.spinonethird, you said that from my description I'm not really using the db and maybe I can roll my own specialized data structure. You and bd seem to think alike, and with two of you in consensus I'm becoming more attracted to this idea, especially if it has the potential for faster performance than an off-the-shelf database program.AS far as the business details are concerned, I know some people (gsiener, rbanffy and perhaps others) are curious but I'm not at liberty to discuss the details at this time. When I eventually spill the beans it will probably be after the business launches so we can gain traction and market share before the competition invades our turf ... :)One question before I end this post:Is C the fastest language or the best language to "roll your own" databse solution, or are other languages just as good or better? I'm not up-to-date on the advantages of many programming languages so your thoughts and opinions on this issue might help me to pick the best language for a project like this. Thanks. |
How do Flickr, Youtube and other high traffic websites handle file uploads? | ars: You are mistaken and it's not true. If your site is php, let php handle the uploads. |
How do Flickr, Youtube and other high traffic websites handle file uploads? | staunch: One really easy way that works extremely well is to use an old school CGI script to handle the upload. It will die as soon as the upload process is finished, which keeps things very self contained and clean.The most important thing (as others have noted) is that you do processing asynchronously. Get the file on the server, queue it (however simply), and then process the uploads in however big of batches your machine(s) can handle optimally. 99% of the time you're going to want ffmpeg for videos and ImageMagick for images. |
How do Flickr, Youtube and other high traffic websites handle file uploads? | mlLK: Here are some interesting stats in how much data 4chan handles on any given day while the upstream might not quite compare to something of the likes of Flickr or Youtube, but from what I can tell while software is certainly important it is your hardware that will make or break whether or not your site can handle such volume.4chan is currently powered by seven servers (five content, two database). We are colocated on a full 500mbps Global Crossing connection, allowing us to push over 5TB (5,000GB) of data per day [Image: http://content.4chan.org/img/traffic.feb5-12.png] |
How Much Can I Scrape? | ca98am79: what kind of data do you want? real-time? or end of day? If you just want end of day data, it is simple. Just write a script that collects it in the middle of the night and stores it in your database. I don't think they mind at all if you just do it once a day for all of the stocks - I know people who have been doing it for years. They use this:http://www.gummy-stuff.org/Yahoo-data.htmIf you want real-time data, good luck. It will cost you. |
What's your Twitter rank? | jfarmer: A friend of mine who is an engineer at Google wrote this up. It uses Real Math(tm), so it calculates your TwitterRank by taking into account how you interact with other Twitter users.Mine is 17.2, which surprises me. |
What's your Twitter rank? | brk: Why does it need a password? twitter.grader.com does essentially the same thing, but doesn't require a password. I'm not sure what additional data the author gets by using the Twitter API vs. just looking at the public info (I can't think of any part of my Twitter stats that aren't publicly available) |
How do Flickr, Youtube and other high traffic websites handle file uploads? | ezmobius: Use this nginx module: http://www.grid.net.ru/nginx/upload.en.htmlIt is highly scalable and spools the uploads to disk and does the mime parsing in efficient nginx C code> then once it finishes it just passes some params to your backend processes with the location of the file on disk and you can process it however you want. |
How do I handle a million hits a minute? | iigs: Sorry for replying to the root, this thread is kind of a mess, with stuff scattered everywhere. Here's what I think:What you're asking to do is very hard for anything beyond simple static page retrievals. Tiny mistakes in schema design (let alone architecture) could make or break your site.Get expert architecture help:If there's no way for you to start small and make mistakes early, your best bet is to find some knowledgeable people, say 3-5 that have professional or large open source scaling experience. Ask them to NDA if you must, but they won't be able to help you unless you can give them all of the details. Solicit their opinions for architecture, and try to reconcile the dissenting opinions about any given technology to see if anyone is off base. Proper architecture could take this from a seven figure hardware expenditure down to a five figure one.Get commercial support for your hardware and software:Once you have a rough idea how you'll do it, you'll want to engage good support from any vendors possible. When you're starting small you can just install Ubuntu, apt-get install mysql and get to work, but at millions of hits per hour you may well start to uncover unusual things involving bad hardware, hardware/kernel, or kernel/userland interaction -- likewise with any DB or other userland software. I would say expect to buy tier 1 hardware, an OS with a support contract, and support for any major software component as well. This isn't a very web 2.0 formula, but you're buying expertise that is focused on your solution.Other things:1) Sharding isn't a magic bullet but depending on how your data is laid out (specifically what data is not local to other data) it could totally save your bacon here. One box (or a standard three tier architecture) probably wouldn't even have a chance.2) SSDs aren't ready yet, IMO. Due to the fact that failure is write cycle count related and not heat/time related, they're not suitable (at this time) for use in RAID systems. I've read reports of SSDs in RAIDs that die in such a tight timespan that you can't get the first one rebuilt before the second one dies.Good luck. It sounds like a really cool project. |
How do Flickr, Youtube and other high traffic websites handle file uploads? | jdavid: so for the most part i don't think i can answer in detail, but in public situations hi5's response has been that they have a dedicated pool of servers running the upload, and they store the file on a static server. once that server fills up, they provision a new one. i will call this the viking image upload process, as its like a viking burial process. each one of those images is then buffered by a CDN.i should point out that hi5 has more photo uploads per day than flickr does. |
How do Flickr, Youtube and other high traffic websites handle file uploads? | eax: one of the Flickr engineers, cal Henderson, wrote a book with a title something like "building scalable websites" that was published by O'reilly. I'm pretty sure he covers that topic. You may be able to get acces online via your public libraries web site (you can in seattle, at least).There are the obvious issues with file uploads, they can take a lot of bandwidth, and disk space, but there are a lot of less obvious problems.1. File uploade take a lot longer than most web requests, both because of the size of the data, and because most client connections download faster than they upload.2. As a result, file upload requests hold server resources longer than other requests. This usually comes down to memory, but there can also be file handle and socket limits. Also, more in the past than in the present day, just the CPU overhead from dealing with lots of open sockets could get to be an issue.3. File uploads often carry a lot of memory overhead. The braindead simple way of handling way fileuploads in PHP, etc ends up buffering the whole file in memory until it the upload is complete. That can really add up. Furthermore, the process handling the upload request has the memory overhead of the PHP (or ruby, or python...) interpreter, and any code and libraries associated with your application. This overhead is carried even though most of that code and data structures are unnecessary for most of the request durration.This memory useage really stacks up when each upload request lives for seconds, or minutes, rather than the milliseconds required for most requests.There are lots of ways to deal with the resource issues. Writing the upload to disk as it arrives is a big improvement. You can go further by having a separate app/server instance that is tuned to minimize the size of each app/interpreter instance is another.There are also ways to take advantage of file upload features built in to a front end webserver (like nginx) to buffer the whole upload to disk before your app has to get involved. Not to mention the amazon examples mentioned.Turning to a specialized custom file upload server written in Java or C seems like an optimization you undertake if you outgrow the other solutions (including more memory per server, or more servers)Imple |
ASK HN: Google account disabled. Oh what to do? | hbnickcarter: Just had an exact repeat episode…account disabled. I my case its because there was a security issue with Google Checkout…the worthless redundant buying service that I signed up for to get 10% off some gear. What a fool. The Googlesphere is inherently evil and here’s why. They promote trust that lures users with awesome free cloud computing service. I foolishly trusted Google and used Gmail, Docs, and Picassa to safeguard my data. Now, because of some BS Checkout problem, they’ve locked me away from my data. IMAP made the situation far worse, since I now have lost 3-years worth of email data. INSANE. I blasted the cheerleaders at Lifehacker who promote all things Google. I suggest that anyone who suffers a similar fate do the same. The word needs to get out that Google can’t be trusted. |
How do Flickr, Youtube and other high traffic websites handle file uploads? | tzury: If you plan to deploy your client application as an HTTP application then there isn't much you can do other than let your web application handle it (php/python/ruby whateveer).Another option is writing an uploader as a stand-alone application (such as flickr uploader, facebook's iPhoto plug-in and alike)The third option is BAD but still exists on facebook as java applet within the browser.The forth option is to write the client in one (or many) of the browser extenders such as Google Gears, MS Silverlight, Adobe Air/Flex/Flash to do this (look also at http://www.jnext.org and http://www.google.com/search?q=flash+uploader).All these 3 can be implemented at the server side in whichever language you choose. |
Review my startup -- Wedding Registry Direct | pclark: tell us about itwhere'd you get the idea, why?how do you hope to make money? |
Review my startup -- Wedding Registry Direct | pclark: - "Bringing wedding vendors together with bridal guests" thats too dark on the heading image- I really don't like the stock background image on this page - http://is.gd/7bID- Is the business type page a joke? There at least needs to be an "other" option- The tour should be a video, rather than a manual slide show- Your privacy policy doesn't exists, that removes all credibility of your site.- can you make the fonts of your menus a bit larger by default?- I'd put your "contact us" option at the top, ideally top right.- can you make your title tag show me where I am? thats nice usability.- alt tags on linksI like this site, its nice and clear, with obviously quite a lot of 37 signals ala design. Not a bad thing.The pages feel a little too noisy, try removing the stock image background. Also be wary of lighter than black colours on darker than white backgrounds.I love the "Access New Revenue Today" - lovely and clear and really grabs me. |
Review our startup - BeatMyPrice.com | unalone: Is the tag cloud really necessary? I agree with the people who've said that you make good designs, but tag clouds are inherently ugly. And I genuinely don't know: do they get many hits? |
What's your Twitter rank? | adoyle: Remember the study where people gave away their passwords for a candy bar? http://www.theregister.co.uk/2004/04/20/password_surveys/That would clearly not work with the HN community. Instead, we need a geeky screen with a disclaimer and the promise of learning how we rank compared to others.What's next? gmailrank? |
How do I handle a million hits a minute? | scumola: It all depends on how "lean and mean" your app and the data is. If you do everything in javascript up-front and use AJAX to move data back and forth, then you reduce the pure bandwidth except for the initial javascript load (which you could cache on Akamai or something).Some math: Sustained (theoretical maximums) with a 1Gbps link gives you 1M hits per minute at 8053 bytes per hit max. However, you're going to get less than this in reality. Figure 80% of this (6442 bytes) is what you can expect on average. A Gigabit pipe is also big bucks.You'll need > 1 web server to handle the CPU load to handle 1M hits per minute, so you'll need a load-balancer that can handle Gigabit (Major $$$ already - we're talking in the $50k-$100k just for a single load-balancer that can handle this kind of traffic and most enterprise load-balancers max-out at one million 'sessions' (open connections)), so you may need to address multiple load-balancers running in an active/active configuration and > 1Gbps pipe if you are moving > 8k per 'hit'. Keep in mind that there is connection time/bandwidth overhead and http packet overhead as well. If you could get away from the browser/http model and move to more of a client/server application, you could optimize better and make better use of your bandwidth. Just a thought. Switching to a UDP protocol could save you bandwidth and make things a little faster network-wise also.To be speedy, your app will have to live without a traditional DB. Put the data in memory (easy to do with memcache across several machines) and make sure that your app is quick and avoid using anything else that would be sub-optimal. Use a lightweight web server, or just get your app to take the http requests itself. C is a good choice for squeezing as much as you can out of your CPU. Write key/value pairs to memcache as your DB, and you should end up with a 10ms turnaround with your app if there is not too much calculation to be done (no sorts, no big/nested loops, ...). Perhaps an apache module for your app would be in order here? Get a few machines maxxed-out with this configuration (need 6+ just as an initial guess, depending on the load-balancer and network latency to serve all of those requests in a second), and it's do-able. Buy reliable machines (DELL or the like) and reliable network equipment since you'll be pushing the hardware to their limit and they'll be running hot.Instead of writing to a conventional database, you can write to a log file or syslog to another machine (1M writes a minutes shouldn't be that bad) and reconstruct the data to a database when the heavy usage slows down (process the log files offline or something on a different machine).This is not a small task and will require some tight code and big $$$ up-front, but it's doable. Your Co-location bill will be large. |
How do I handle a million hits a minute? | crabapple: lots of hardwarethere are a few best practices to scaling, you can find those archived at places like highscalability etc...but in the end you just need hardware. hardware buys you fault tolerance, test environments, sharding, blah blah blah. if you are looking to meet this traffic on your own box or on a vm-hoster, forget it |
How do I handle a million hits a minute? | lallysingh: First, relax.Look, you've got ~4 GHz processors available, and you're worried about a million hits a minute. That's nearly 240k clock cycles per hit.Next, you've got 60 +/- 10ms of latency between you and your clients (unless there are details here you aren't covering, but no matter). You don't really have to respond to users terribly fast, as lots of your responsiveness would be lost over jitter/lag.A single dell box could handle your load, if you wrote the whole thing in C on a modern quad-core desktop.Ignore all the standard Apache/SQL/networking stuff until later until you know what you need. Get the core requirements up first, then add on the load of the side-stuff secondly.E.g. doing all the heavy lifting in a private set of cpu cores and then providing a shared-mem plugin to apache may be enough for you. Save a CPU core to snapshot your data.So, for advice:1. Ignore database bullshit. You don't need it, it won't help. If you want a DB for other purposes, fine. A snapshot process writing to your DB is fine, just don't put it in the critical path.2. Build a load simulator. A raw mode that just sends over the handful of bytes, and a cooked mode that bothers to printf' a GET request.3. Start with a reasonable prototype, and work your way to something performant. Hell, you can probably do it in java if you don't mind buying a few more CPU cores.4. Integrate as you need for the rest of your requirements. For example, have another box serve the rest of your webapp, and dedicate a stripped down apache box with a custom module for this stuff.In essence, I'm telling you to treat it as a very smallish HPC problem, instead of some sort of nightmare webapp problem. It fits better, and suddenly you have lots of people/knowledge/COTS equipment available to you. |
Solid MIME parser library in C or C++? | wmf: Look into Evolution and Thunderbird; both are written in C and are being maintained. |
How do I handle a million hits a minute? | spc476: Most of the responses here focused on the disk/data aspect. What about the actual network? IPv4/TCP has a minimum overhead of 40 bytes per packet, with a seven packet overhead for the connection itself (three for making a connection, four for closing the connection). Adding another packet for the actual data (with say, a payload of only 20 bytes) and that's about 220 incoming bytes across 5 packets (two from the client for the connection, one for the data, and two for closing the connection---I'm concentrating on the incoming side as outgoing should be similar). Multiply 220 bytes by 10 to get a rough estimate of bits per second (close enough and makes the math a bit easier and if we're a bit high here, that's better than being too low) times the number of connections per second (16,667) gives us around 37Mbps---a T3 basically.Now, to handle a potential 100,000 users at the same time, that's 6 T3s (12 if you want to handle 200,000), and that's with only 20 bytes of actual data being shipped between client and server. Just the networking side is going to be expensive, so I hope the idea is a good one.One the data storage side, I worked on a project (http://www.x-grey.com) where I had to potentially store a bunch of records. I ended up storing them all in memory in an array. When a new record came in, I would insert it into the proper spot. In theory, I was able to handle about 6,000 requests per second on a 2.6GHz machine (although I found it rather difficult to actually test it properly, which is why I say in theory---in practice it may be more or less, but for what I was doing, that was way more than good enough). So yes, I would recommend keeping everything in memory and avoid the disk as much as possible. But then you have issues of coherency if you have multiple threads/machines that need to update this structure (in my case, that 6,000 is ONE CPU doing the work; I sidestepped the whole concurrency issue by ignoring it). |
How do Flickr, Youtube and other high traffic websites handle file uploads? | dhotson: http://code.flickr.com/svn/trunk/uploadr/ |
How do I handle a million hits a minute? | midwestward: You biggest problem by far is getting a million hits per minute. Getting a million hits at all would be a fairly rare accomplishment. |
How do I handle a million hits a minute? | acangiano: As far as the database goes, with decent hardware DB2 9.5 will have you covered [1]. And, should you become truly huge, it can exceed your requirements by a long shot, provided that some serious hardware is employed [2].[1] http://antoniocangiano.com/2008/11/04/benchmarking-db2-purex...
[2] http://www.tpc.org/tpcc/results/tpcc_result_detail.asp?id=10... |
How to I process one million hits a minute? | owkaye: Business Detail #1: IT'S A LIVE ONLINE AUCTIONI'm not at liberty to discuss the industries or specific applications in which this platform will be used so please don't ask. But I can tell you that the platform is structured like a live online auction with some unusual / unique requirements:Each auction will run for about an hour, beginning and ending at a pre-specified time. In the specific industry where we will launch this platform we can absolutely limit the number of bidders in every auction to 200,000 concurrent bidders, and we expect only about 100,000 or so bidders to be concurrently logged in and bidding. |
How to I process one million hits a minute? | owkaye: Business Detail #2 - LIMITED NUMBER OF BIDSEvery bidder gets the same limited number of bids in a particular auction, for example 200 bids each. This means the server software must "count the bids" for each bidder and disable bidding for every person who uses all his/her bids prior to the end of the auction. |
How to I process one million hits a minute? | owkaye: Business Detail #3: ONE MILLION HITS A MINUTE MAXEach bidder who remains online and participates during the auction may submit as many as 5 new bids per minute, thus requiring the server to "deal with" as many as one million hits a minute.The auction structure is atypical because the the highest unique bid at auction's end wins, and the bidding range is limited to 0-10% of the item's true market value. |
How to I process one million hits a minute? | owkaye: Business Detail #4: CLOSED BIDDING WITH LIVE STATUS REPORTSEvery bid is a "closed bid" which means no other bidders know its actual value. Bidders are given the "status" or "position" of their bids in relation to the current unique high bid, but they never see the values of other people's bids.I must send the answers to these two questions to each bidder every time he posts a new bid:-- Is his new bid unique or has someone else posted the same bid value previously?-- Is his new bid > = < the currently unique high bid?If his last bid = the currently unique high bid, the bidder is in an enviable position because he is currently winning the auction. Other bidders must bid the same exact amount (to the penny) in order to cause him to lose his position as the unique high bidder. When this happens the unique high bid value must be determined by the software again, and with 100,000+ bidders is is likely that someone else will become the new unique high bidder.If his last bid is not unique or if it is > or < the currently unique high bid, the bidder must post another bid which he hopes will become the new unique high bid. But he does not know the value of other people's bids, all he knows is the value of his own bids and their position relative to the current unique high bid. He must therefore use intelligence, strategy, and quick thinking (game theory?) to determine the value of his next bid, and the one after that, etc., until he succeeds in becoming the new unique high bidder. |
How to I process one million hits a minute? | owkaye: Tech Issue #1 -- WRITING DATA TO DISKI've always used a database to accomplish tasks like these before, but it seems that what I really need is a faster (RAM-based) way to receive and store every bid value received during the hour-long auction -- along with the bidder's unique ID so I know who made each bid.Once the auction has ended I can write this data to disk, but there's no need to write to disk during the auction itself -- because if the system screws up I'll have to re-run the auction anyways, and in this case the old bids can be deleted. |
How to I process one million hits a minute? | owkaye: Tech Issue #2 - CAN JAVASCRIPT SPEEDS THINGS UP?If I understand correctly, some have said Javascript in the browser can reduce or eliminate the HTTP overhead and dramatically reduce both data transfer and bandwidth requirements. Is this true?I don't know how to do this but I welcome a simple explanation that illustrates how it might work. Right now I'm thinking that HTTP headers still need to be sent upon each request and they will use much more bandwidth than the data itself.Is the ultimate solution to find a way to transfer data without constantly opening and closing HTTP connections? |
How to I process one million hits a minute? | owkaye: Tech Issue #3 - KEEP-ALIVE OR BROWSER PLUG-INS?If a connection can be opened and remain open for the entire hour-long auction this might dramatically reduce the overhead of individual HTTP connections. Is something like this possible with as many as 200,000 bidders? If so, how many (and what kind) of HTTP servers might I need to maintain this many open connections?If this might require a FireFox plugin on the client end it is theoretically possible to make this a requirement. Obviously I prefer to avoid this requirement but if nothing else works this may be a functional alternative. |
How to I process one million hits a minute? | owkaye: Tech Issue #4 - BATCH PROCESSING?I've never done batch processing before, but some have suggested it, and I don't understand how it works just yet. My current concept is that I would receive (for example) 15,000 bids during a one-second period and I would store them in RAM as they are received. Then at the end of this one-second period I would have the software process the batch -- which probably means updating the position of the current unique high bid, determining the status of each of the 15,000 bids in relation to this new position, and creating and delivering 15,000 unique HTTP responses. Is this description totally wrong, or is this somewhat close to the way "batch processing" might work for me? |
How to I process one million hits a minute? | owkaye: TOPIC: CUSTOM DATABASE STRUCTURE IN RAM?ig1 said: "I've worked on a number of high volume systems (million+ client interactions/minute), and you don't want a conventional database. Either use a custom data-structure to keep it in memory (even if it's across multiple machines) or if you really want to use a database use one thats designed for that kind of usage (think tickerplant databases, kx, etc.)"my reply: Others have suggested similar alternatives to using a traditional database program, and right now I'm tempted to learn more about doing this -- unless someone can tell me why I shouldn't?By the way, I've never heard of ticker plant db's before but I'll look into them if others think they may provide an even better soution than a custom coded C application. For those of you who have used ticker plans db's, what do you think of their use in my situation? |
How to I process one million hits a minute? | owkaye: Topic: MY EMAIL ADDRESS, TOP OF THE PAGElallysingh said: "If you're doing something game-ish, talk to me privately. Scalability of video games is my phd topic."my reply: It certainly seems like I'm doing something 'game-ish' although it's not likely to be used by the typical gamer. I've posted my email at the top of the page so please send me an email and we can discuss the details privately. Others can email me privately as well to discuss this or other topics, thanks. |
How to I process one million hits a minute? | owkaye: Topic: ONE OR MORE SERVERS FOR THE DATA?ig1 said: "Figure out how to partition your data/algorithms so you can split it across multiple machines..."but lallysingh said: "A single dell box could handle your load if you wrote the whole thing in C on a modern quad-core desktop."my reply: If I'm not mistaken you guys have a difference of opinion here regarding the ability of the data to be processed on a single machine. Can you expand on the reasons for your apparent difference of opinions? |
How to I process one million hits a minute? | owkaye: Topic: LOAD TESTING DURING DEVELOPMENTowkaye (that's me) said: "This is not a system I can grow into, it must be capable of this performance from the very beginning."then ericb said: "I would suggest you load test extensively and make load testing a part of your development process from the get-go. Initially, I would load test to evaluate approaches and estimate hardware needs.and lallysingh said: "Build a load simulator. A raw mode that just sends over the handful of bytes, and a cooked mode that bothers to printf' a GET request."my reply: I don't know how to do this (or C coding or javascript coding) personally but thanks for pointing out these technical needs and more. Your suggestions will be put to good use when I have a complete concept of the basic requirements, and I can "fill in the blanks" by hiring programmers and buying equipment as I go along. |
How to I process one million hits a minute? | owkaye: Topic: GENERAL PLAN OF ACTIONlallysingh said: "So, for advice:1. Ignore database bullshit. You don't need it, it won't help. If you want a DB for other purposes, fine. A snapshot process writing to your DB is fine, just don't put it in the critical path.2. Build a load simulator. A raw mode that just sends over the handful of bytes, and a cooked mode that bothers to printf' a GET request.3. Start with a reasonable prototype, and work your way to something performant. Hell, you can probably do it in java if you don't mind buying a few more CPU cores.4. Integrate as you need for the rest of your requirements. For example, have another box serve the rest of your webapp, and dedicate a stripped down apache box with a custom module for this stuff.In essence, I'm telling you to treat it as a very smallish HPC problem, instead of some sort of nightmare webapp problem. It fits better, and suddenly you have lots of people/knowledge/COTS equipment available to you."my reply: Thanks for your outline of the best way to approach this project. I don't understand parts of what you said here but I think I get the general idea. Do others have any suggestions to add to this, or to change it? |
How do I handle a million hits a minute? | owkaye: NOTICE ! THIS THREAD IS BEING MOVED !!!I'M GOING TO START A NEW THREAD TO CARRY ON WHERE THIS ONE LEAVES OFF SINCE THINGS HAVE BECOME SUCH A MESS IN THIS THREAD ... AND I'M SHOUTING TO MAKE SURE EVERYONE KNOW ABOUT THIS SO THEY CAN FIND AND PARTICIPATE IN THE NEW, CLEANER AND BETTER ORGANIZED THREAD.Thanks and sorry for shouting, here's the new thread URL:http://news.ycombinator.com/item?id=362810 |
How to I process one million hits a minute? | alecco: You already posted this and you are making comments from what looks like your irc logs or comments from the previous post.http://news.ycombinator.com/item?id=361730 |
How to I process one million hits a minute? | alecco: IMHO you would be much better off just finding someone to deal with all this. A contractor or a partner.[Edit: read someone experienced on this] |
Rate my startup (video) | mdolon: Looks very interesting and promising. I'm not sure if I have much use for it just yet but I think I would be easily convinced after seeing more examples of useful apps. I also like the domain name and applaud you both on the immense amount of work I'm sure it took to make the product. |
Rate my startup (video) | jacobscott: What's the technical underpinnings of the stateful stuff? Is it more than multithreading? Ruby specific? Seems neat, but hard to tell if it will really be useful (in possibly the same vein as javascript-based "OS in a browser"s). Can you come up with a killer app? Or reproduce a well known mashup with very few lines of your own code?On the backend, you'll definitely need sophisticated monitoring/resource allocation stuff once you have users, right? |
Rate my startup (video) | guruz: I like it, it's cool :)
Although I can't currently think of a use for it for myself. |
Rate my startup (video) | wesley: Very nice, but can you skip the first step? You have to send "ping" first?If I use a service multiple times, I would already know what input the app needed. For example, "Weather NY". |
Rate my startup (video) | swombat: "It's a platform for creating widgets for the mobile web". Platforms don't make money. A platform is not a business."We don't really know what widgets are." Following on the point above, you're creating a platform (a non-specific, generic "thingy") for creating widgets (non-specific, generic "thingies") for the mobile web (something that's not even really defined properly yet. Non-specific businesses are almost without exception failures. Only huge businesses like Sun and Microsoft and Amazon can afford to release "platforms" without going broke.Here's the most valuable advice you can get at this point if you want to build a start-up (that's a business, meaning it needs a path to making money) is:Find one or two specific customers who have a need that this fills, and fill it, and get them to pay for it. Specific wins the day.Since you've built a platform, you may well have to go one step closer to customers before you can help them. Design an application that people will be willing to pay for, using your (admittedly very cool) platform. Then, once you have your app and you have profits, over time you might want to open up the platform to other people.Repeat once more: a business is an entity that makes money. A platform is not a business. |
Rate my startup (video) | alaskamiller: You've made a kind of Automator software (http://en.wikipedia.org/wiki/Automator_(software)) for the web. This is pretty nifty! With more integration with services it could actually be pretty useful.Don't know how you can make money though :( |
Rate my startup (video) | mikeyur: Looks pretty sweet, I could probably think of some way to use it :) |
Rate my startup (video) | rgrieselhuber: I think this is actually really cool. I agree with the other commenters that you're probably a long way from this being a "startup" (at least in the post-Oct2008 world that would like to see some business model), but you are certainly doing some interesting things.How you turn this into a business is another problem but I think you will probably see people picking up on the ideas here and rapidly creating their own innovations. That might seem like a bad thing, but I think it's good because it means you've come up with something very powerful.Probably the most interesting thing to me is the subset / scripting language that makes interacting with various services and providing the backend infrastructure for performing that automation.Good luck! |
Rate my startup (video) | davidw: I don't have time to watch videos, sorry.Meaning: I'll happily take a look at your site, and have many times in the past here. What I don't want to do is sit through a video that I can't jump around, explore, and so on, like I might with a site, or a brief article. |
How to get Ads fit the Content? | tok9: I'd like some info on this too. |
Alternatives to Y Combinator? | davidw: You could do cambridge / mountain view for the three months that you're required to be there, then leave - I don't think there's anything in the rules against that. |
Alternatives to Y Combinator? | beaudeal: My first piece of advice would be this: don't bank on YC funding. Sure, it would be great if you got accepted, but that shouldn't stop you from pursuing your idea. Especially if it is still just an idea - work on it at home, in your spare time, and then evaluate your options once you have done everything you possibly can do on your own.If you decide that you need outside funding, and even Mountain View is too far, then your options for YC (or a similar program) become more limited. The next YC round will be in Cambridge, but there is TechStars in Boulder, Colorado which is obviously much closer, but may still be too far. The only other programs I have heard of are in the Philly / DC areas.Like you said though, there are more traditional routes. A Google search for 'arizona angel network' brings up over 2m hits. My point is that, I'm sure there are options, you just need to look for them, and then evaluate.First things first, though: start hacking! |
Alternatives to Y Combinator? | cperciva: So, is there a similar program but without the move? [...] The biggest attraction of YC to me is the cumulative experience and knowledge of the founders, and I worry that I wouldn't get that with a standard VC path.How do you expect to get a "similar program" without moving? The cumulative experience and knowledge of YCers isn't something which can be conveyed remotely.My google-fu is failing at the moment, but I remember seeing a comment from pg here a while ago to the extent that YC once tried funding a team remotely, but it didn't work due to the lack of frequent contact. Don't look at YC's move-across-the-country requirement as a roadblock they're putting up -- look at it as them trying to maximize the chance that the companies they fund will succeed.(That said, I submitted a YC application in spite of not being able to move -- but I wasn't really expecting to be accepted, and if I had been accepted, it would have taken sitting down and trying to figure out exactly how YC could help a non-local startup before I would have accepted.) |
Alternatives to Y Combinator? | markessien: You probably have a day job, right? Put aside a certain amount everymonth and hire people on elance to create your product for you. In a few months it will be done and you can test it on the market. It's risk free, and it's the sensible solution for someone with a wife and children. |
Alternatives to Y Combinator? | pragmatic: Why do you need seed/angel funding. Why don't you bootstrap it? Work on it in your spare time? And don't give me excuses about a young daughter. I did it with a young son while working full time. It's about desire, not location. Don't buy into the hype.I'm not convinced YC is a good thing. What about the group think problem? Hey everybody let's build sites then [..magic..] profit!If it's a good idea slow and steady will do just fine. |
How to I process one million hits a minute? | smoody: Have you checked out memcachedb? It's a wrapper around the berkleydb that uses the memcahed protocol (so support is already available in your language of choice). They're showing benchmark stats that meet your requirements (> 1MM writes per minute on a single box, obviously more boxes = more writes): http://memcachedb.org/benchmark.htmlIt's simply key-value pair storage but it sounds like that might be all you need. |
Alternatives to Y Combinator? | theklub: Anyone mentioned Vencorps.com? If no one has I'd check it out. |
How to get Ads fit the Content? | guruz: Take a look at
https://www.google.com/adsense/support/bin/answer.py?answer=...
Maybe it helps. |
Rate my startup (video) | ionrock: I agree with most folks that it does really solve a problem, but in an effort to try and be constructive, one potential problem are forms on mobile devices. It is very difficult to port web applications or forms to mobile platforms because the display area is so small. Creating a conversational interface via chat might be a good angle to creating usable, yet complex, forms on mobile devices.One good example might be ordering a pizza at a party. It is loud and you don't know the address. You can start texting a pizza place your order and let some other aspect of the application handle the GPS position.Anyway, it looks like you guys have been doing a good job so if this idea makes you millions or something feel free to send an email or something ;)http://ionrock.org |
Review our startup - BeatMyPrice.com | truebosko: Wow, very cool. I just tried it on a product on our website (We are a Canadian retailer) and it worked!I'm curious how are you crawling the web? As a small business, it doesn't find us unless I actually enter our url (even though we have decent rankings in Google for many of our products)This would be neat to make as an API so retailers like myself could plug it into our stores websites. |
Alternatives to Y Combinator? | rsmerritt: There is a new Y-Combinator style accelerator forming in Arizona called Tekcelerate (www.tekcelerate.com). |
Alternatives to Y Combinator? | sanj: Consider the fbFund.I'm happy to chat about it. I've written a little about it too: http://blog.luckycal.com/?p=11 |
Mailing Newsletters | makecheck: An opt-in mailing list system such as GNU mailman would reach the people who want to be reached and give them plenty of options (e.g. individual mails versus digest form, the ability to turn it off temporarily for vacation, etc.).I would strongly recommend you not automatically E-mail your users on a weekly frequency! If you must do this, make it infrequent (monthly, no sooner) and make sure you include genuinely useful information. A good example is LinkedIn, which includes information that I can definitely use (new developments among my directly connected peers), and they only send it if I haven't visited the site in about a month. |
Mailing Newsletters | jmah: http://www.aweber.com/ (not that I've used it) |
Please review my startup | vaksel: This looks pretty much like a knock off of google financeAs far as site wise
-too much space on the bottom
-newstrends are pretty useless for terms, since you get results like you have now "EPS" etc, I'd replace that with newstrends for actual companies. |
Please review my startup | blurry: Love the UI but feel a little ambiguous about the basic principle behind it. Would love to hear others' opinions about it though... I could be wrong on this one.As I understand, you basically attempt to correlate stock movement with two major components: (1) market/sector movement, and (2) news.The first one is easy. Market/sector goes up, stocks go up, market/sector goes down, stocks go down (in absence of major news, more on which below). I doubt you would have created the website to prove the obvious, so I'll leave this one alone.The second one is where I see the major problem. The problem being that NEWS DO NOT MOVE STOCKS. You'd think they would. After all, good earnings report, stock goes up, bad earnings report, stock goes down, right? Except that's not what happens. Granted I have limited trading experience (selling/buying/sometimes shorting several times a month at the most) but I've seen enough to conclude that there is no meaningful correlation between news and stock movement.I just tracked several stocks that I happen to follow where I have a reasonable idea as to why they've moved in the last year. Your chart only proved the above. News were well presented along the spikes/dips (again, the UI is awesome) but the clear presentation only served to show that no correlation exists. (To clarify, by news I mean anything and everything, e.g. earnings reports, analysts upgrades/downgrades, company announcements, actual news, etc; and by no correlation I mean good/bad news accompanied spikes/dips equally.)Anyone care to argue with me? Prove me wrong! |
Please review my startup | Aries09: this site is great as a tool
i think you are wrong by calling it a google clone
also with all that's happening now on wallst i get a chuckle checking it out |
Please review my startup | davidw: It looks pretty, and looks kind of interesting, but where's the 'there'? Like vaksel notes... it's pretty similar to Google Finance - what's the compelling reason to keep coming back to your site? |
Please review my startup | mstefff: Doesn't Google Finance do this already?I don't understand the use of the "index"...The logo and interface looks a little '5 years ago'.. |
Please review my startup | steveplace: A couple notes:- It might be a good idea to note stock:market divergences on your chart- Don't know where your parsing your news, but you might want to incorporate a filter for SEC filings and PR releases along with a way to eliminate the fluff/SEO (Motley Fool comes to mind)It seems (as blurry pointed out) that you are assuming that news leads/lags a stock. There's also a whiff of efficient market theory in there, and at the expense of pissing off half of HN, the markets aren't efficient.Other than that, I find a hard time finding the difference between what you have and other tools I use. |
Rate my startup (video) | decadentcactus: I like it, I'd probably at least play around with it or integrate it if I found a use. Great work! |
Please review my startup | tptacek: Are there other things you can chart and correlate news stories to, besides the stock market?Can you put in the extra effort to chart indicators, like VIX and TED and Case-Schiller? CSCO and MSFT stock charts might be tapped out, but this might become useful if it acts as a portal to lesser-known stats. |
Please review my startup | cabalamat: Here are my observations for the page http://whomovedmystock.com/GOOG1. it looks pretty2. when I move the bar at the bottom of the chart, the applet keeps reloading. This is annoying. Is it intentional or a bug?3. the bit at the top left where it shows a pie chart of Google's "Market index allocation" -- how is this calculated?4. On the chart, it's not obvious to me what the blue line "market index" means |
Please review my startup | sahaj: reading the title, i thought, oh cool, this will tell me which investment firm(s) are buying/selling a particular stock.as others have pointed out, this doesn't provide anything more than what google finance already does. |
Please review my startup | volida: the graph doesn't look very professional/serious because of the curves in the change of the direction. |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.