mirror of
				https://github.com/donnemartin/system-design-primer.git
				synced 2025-10-31 08:12:31 +03:00 
			
		
		
		
	Fix typos in Twitter and web crawler exercises (#438)
This commit is contained in:
		| @@ -26,7 +26,7 @@ Without an interviewer to address clarifying questions, we'll define some use ca | |||||||
| #### Out of scope | #### Out of scope | ||||||
|  |  | ||||||
| * **Service** pushes tweets to the Twitter Firehose and other streams | * **Service** pushes tweets to the Twitter Firehose and other streams | ||||||
| * **Service** strips out tweets based on user's visibility settings | * **Service** strips out tweets based on users' visibility settings | ||||||
|     * Hide @reply if the user is not also following the person being replied to |     * Hide @reply if the user is not also following the person being replied to | ||||||
|     * Respect 'hide retweets' setting |     * Respect 'hide retweets' setting | ||||||
| * Analytics | * Analytics | ||||||
| @@ -129,7 +129,7 @@ If our **Memory Cache** is Redis, we could use a native Redis list with the foll | |||||||
| | tweet_id  user_id  meta   | tweet_id  user_id  meta   | tweet_id  user_id  meta   | | | tweet_id  user_id  meta   | tweet_id  user_id  meta   | tweet_id  user_id  meta   | | ||||||
| ``` | ``` | ||||||
|  |  | ||||||
| The new tweet would be placed in the **Memory Cache**, which populates user's home timeline (activity from people the user is following). | The new tweet would be placed in the **Memory Cache**, which populates the user's home timeline (activity from people the user is following). | ||||||
|  |  | ||||||
| We'll use a public [**REST API**](https://github.com/donnemartin/system-design-primer#representational-state-transfer-rest): | We'll use a public [**REST API**](https://github.com/donnemartin/system-design-primer#representational-state-transfer-rest): | ||||||
|  |  | ||||||
|   | |||||||
| @@ -77,7 +77,7 @@ Handy conversion guide: | |||||||
|  |  | ||||||
| ### Use case: Service crawls a list of urls | ### Use case: Service crawls a list of urls | ||||||
|  |  | ||||||
| We'll assume we have an initial list of `links_to_crawl` ranked initially based on overall site popularity.  If this is not a reasonable assumption, we can seed the crawler with popular sites that link to outside content such as [Yahoo](https://www.yahoo.com/), [DMOZ](http://www.dmoz.org/), etc | We'll assume we have an initial list of `links_to_crawl` ranked initially based on overall site popularity.  If this is not a reasonable assumption, we can seed the crawler with popular sites that link to outside content such as [Yahoo](https://www.yahoo.com/), [DMOZ](http://www.dmoz.org/), etc. | ||||||
|  |  | ||||||
| We'll use a table `crawled_links` to store processed links and their page signatures. | We'll use a table `crawled_links` to store processed links and their page signatures. | ||||||
|  |  | ||||||
|   | |||||||
		Reference in New Issue
	
	Block a user
	 Agade09
					Agade09