This video is soo bad, you get easier to understand information just by reading the paper and articles.
@prashantv21702 күн бұрын
what is the limitation of displaying number of columns and export number of records from a tabular report from the Databricks BI? can you show 500+ columns and export 1 million records from the report?
@hcubill3 күн бұрын
Horrible sound, would be great to check guests sound before recording
@neelbanerjee78753 күн бұрын
Can we do this publishing in a automated way? programatically or using a scheduling?
@dahof27894 күн бұрын
Wow!! Using your laptop mic for this?! Are you kidding? 😂😂😂
@pavanbarnala5 күн бұрын
like the tone of this video :-)
@TwanManders-n7r5 күн бұрын
why talk so fast?
@voxdiary6 күн бұрын
these are solid tips. thank you for revealing the framework
@deathseal98446 күн бұрын
Hi, amazing video but where is github link. which host is talking in video.
@Wedzol6 күн бұрын
can you access the data with a 3rd party BI tool such as PowerBi or Tableau?
@balajikasiraj6 күн бұрын
Doesn't the "join" condition execute before the "where" ? Therefore the filter should be post the join and not pre-join in scan B at 6:32
@SeaWhiz7 күн бұрын
Great demo of Databricks' clean rooms! As someone with interest in fintech, I really appreciate how this enables secure data collaboration between banks without compromising customer privacy. The ability to run AI models across federated data while maintaining strict data governance is impressive. Also love how it works seamlessly across different clouds and platforms. This is exactly what the financial industry needs for better fraud detection.
@alvine88117 күн бұрын
Thanks for being realistic
@evilapple34278 күн бұрын
Is there any particular advantage for using the transformer flavor instead of deploying LLM as a custom pyfunc? Are there any optimization such as flash attention or page attention for speeding up inference applied by the transformer flavor? Or does it improve loading speed of the model?
@YanyanDog9 күн бұрын
Thanks for the great video! I've enabled Predictive Optimisation for my unity catalog a couple days ago. But nothing happens, I cannot see any records in the predictive_optimization_operations_history schema under system catalog. Does it mean Databricks believes it is NOT worth optimising my tables because of the cost?
@Databricks4 күн бұрын
Hi Yanyan, although cost is a factor, the other factor is how much you'll benefit from that optimisation. If you have lots of tables that don't get read that much, then it might decide it's not worth it. It can be quite conservative. I turned it on in my teams demo workspace, but because we don't do 'real' sized data, it didn't do anything. At 1:12 I had to use one of the bigger field workspace that has nearly 1000 users in it. No harm in keeping it turned on though - Holly
@YanyanDog3 күн бұрын
@ thanks for your detailed explanation! I just realised that I forgot to enable “storage” schema under “system” catalog. Now I can see the predictive optimisation logs.
@lingxiang619210 күн бұрын
a worked example of using Databricks vector search will be helpful.
@JohnV-e6g10 күн бұрын
Awesome!
@AnkurRoy-k8b10 күн бұрын
Setup up delta-io/delta is barrier due to multiple complain in compilation process, making it very difficult to contribute. 😥😥
@gordeyvasilev10 күн бұрын
👍
@bobbysega383911 күн бұрын
Is there any reason why this couldn't be done in native Terraform? The Databricks provider provides much of this functionality as far as I can tell.. And of course Terraform natively supports lookups, interpolation syntax etc. And as I understand it, DAB is just a wrapper around Terraform. So if we already have expertise in Terraform, and have paid a considerable amount for Terraform Enterprise, this seems like we'd be adding an extra layer on top of stuff we can already do? Or does DAB do something that the Terraform provider cannot?
Why none of videos works for me? is it the location being US is the problem, it just shows the ad and skips to the next video..
@RaviY-o6r12 күн бұрын
Just Like a WOW
@And_Nunes13 күн бұрын
can you make a tutorial on how to build this dashboard, please?
@dhruvpathak185014 күн бұрын
good summary and key points..
@dejavu508514 күн бұрын
Love the energy !!
@criticalnodecapital14 күн бұрын
This is gold
@dejavu508514 күн бұрын
How would you know what is the best algo or best forecast ?
@DebayanKar715 күн бұрын
Genius way to rip off customers
@mitospha15 күн бұрын
So do each of the 4 microservers that replace 1 SQL server, need their own CPUs just to run? Doesn't that mean that it requires more overhead for the same process?
@marcpresume311115 күн бұрын
Can you please share the codes with us?
@marcpresume311116 күн бұрын
Can we access your code?
@rhambo555416 күн бұрын
I get emails all the time from my CEO, they always come with an S3 path too. Meanwhile, in reality...
@RomuloMagalhaesAutoTOPO17 күн бұрын
👏
@asiasowa17 күн бұрын
Great talk :) It would be fantastic to hear more about Localhost config and Spark Connect considerations in the near future :)
@420_gunna17 күн бұрын
Just clicking to say the facial hair goes crazy
@Alex-voini19 күн бұрын
Love the jokes at 5:50 🤣
@phawatmakhaphan132919 күн бұрын
Can I use lakehouse federation to query data from hive table which ran on server if I have hive, hive metastore database (mysql).
@balaji322920 күн бұрын
Hi @Ruifeng Zheng, AFAIK , there is no Java Client for Spark Connect in the Spark 3.5.x . Will it be available in Spark 4.0 GA?
@jithujosekokken20 күн бұрын
the best detailed explanation!
@gauravbhure703320 күн бұрын
Stored Procedure is not working. #Databricks
@vidurbhardwaj400120 күн бұрын
Nice knowledge sharing session. Can you please share the ppt you are referring
@balaji322920 күн бұрын
Hi @Hyukjin Kwon, Please help me crack this error. When trying to programmatically (Python) access the Spark Connect Server, doing a simple dataframe query throws the error. I could ensure there is no connection issue. Python Logs: pyspark.errors.exceptions.connect.SparkConnectGrpcException: <_MultiThreadedRendezvous of RPC that terminated with: status = StatusCode.CANCELLED details = "Received RST_STREAM with error code 8" debug_error_string = "UNKNOWN:Error received from peer {grpc_message:"Received RST_STREAM with error code 8", grpc_status:1, created_time:"2025-01-12T08:45:47.1424932+00:00"}" > Spark Connect Server Error: org.sparkproject.connect.grpc.StatusRuntimeException: INTERNAL: Encountered end-of-stream mid-frame
@ns_the_one21 күн бұрын
just a feedback. Use more meaningful images rather than talk-telling.