Kafka -> Loki -> S3 If you're looking for an existing solution :)
@interviewpen10 ай бұрын
Yep, S3 does a lot of the things discussed here behind the scenes. Thanks for watching!
@lunaxiao99978 ай бұрын
great video,very clear
@interviewpen8 ай бұрын
Thanks!
@developerjas10 ай бұрын
Great Video man! Would how would you go about designing the data ingestion part?
@interviewpen10 ай бұрын
Great point! There’s a lot that goes into ingesting logs while optimizing network performance and maintaining context. Check out our full video on monitoring systems on interviewpen.com :)
@sahanahunashikatti393510 ай бұрын
😊😊 ok 0@@interviewpen
@supragya80556 ай бұрын
i dont understand , if under same bucket lets say for (2021-2022) we have multiple nodes , how are reads any faster ? for the same bucket logs will be distrubuted across servers and still need to be queried across servers which is slow . Bucketing didnt help in improving read performace , is my understanding .
@interviewpen6 ай бұрын
Yes, sharding improves write performance at the expense of query latency (unless we shard by something more clever!). However, we can still handle a high throughput of reads. This latency vs throughput problem is a common tradeoff with large-scale systems! Hope that helps :)
@sahanagn44857 ай бұрын
Great video!!! Please slow down the speed of video as someone new to topic its bit fast to grasp the concept.
@interviewpen7 ай бұрын
Ok, noted!
@GoofGoof-cs6ny6 ай бұрын
So in 2018 every service was writing logs to node 3, didn't we went back to bad write complexity by doing bucketing?
@interviewpen6 ай бұрын
Yep, bucketing makes query performance better, so we introduce sharding as well to distribute writes within a bucket.
@michatobera604910 ай бұрын
Great video
@interviewpen10 ай бұрын
Thanks!
@didimuschandra668010 ай бұрын
Greatt video!! thanks! but, can you create video to develop Effective and efficient Ticketing System?
@interviewpen10 ай бұрын
Sure, we'll add it to the backlog. Thanks for watching!
@prakharsrivastava66443 ай бұрын
I love the cute computer in the background
@interviewpen2 ай бұрын
Thank you :)
@weidada10 ай бұрын
Suppose every two years, it ingest 2PB and migrate 1PB, how could three sets be enough to cycle after 12 years?
@interviewpen10 ай бұрын
Great question! At any given time, we have three "hot" nodes--two are migrating data to cold storage and one is ingesting new data. We only showed one cold storage node in the example, but we would need at least 2 to make this work long-term. Hope that helps!
@ankushraj35995 ай бұрын
Why not use Kafka for high through put?
@interviewpen5 ай бұрын
Kafka is an event streaming platform, so it wouldn't solve any of the log storage problems we're addressing here. But if you have any thoughts on how to incorporate it, feel free to share!
@RaushanKumar-co3wj5 ай бұрын
@@interviewpen Use kafka stream + cassandra . process the event through consumers and save inside a Hbase db for analytics .