Transcript for:
Databricks SQL Demo

databrick sql offers a scalable fully managed interactive query engine for all bi and sql workloads that outperforms leading data warehouses on a price performance basis it offers best-in-class open connectors to the bi and reporting tools you already use like power bi looker and tableau as well as a built-in sql ide by the end of this demo we'll show you how easy it is to deploy manage and administer data bricks sql to enable self-serve analytics directly on the data bricks lake house platform so let's get started the first thing we want to do as administrators is to invite other users to our data bricks workspace we can add new users by email like we're showing here or databrick sql can leverage your existing user roles and groups by automatically syncing with azure active directory awsim octa single sign-on or your cloud identity provider of choice once users have access to databricks sql they'll be able to log into the lake house and we can start giving them permissions to access data and run queries now that we've invited some users let's return to the data bricks home page from here we can jump over to data bricks sql by clicking the sidebar and selecting the data brick sql lens you'll see the brand new data brick sql guided onboarding experience in the panel on the left hand side go ahead and click configure data access as the first step in our journey from this screen in the sql admin console you can turn on databricks serverless sql serverless sql endpoints provide instant compute for all your bi and sql workloads with minimal management required down in the data security section we can add our data lake storage credentials into databrick sql and configure it to enable users of our endpoint to access data in the lake house next let's jump over to the data explorer to view some data and manage permissions from here data analysts and users can discover all the data sets in the lake house browse schemas and see sample data let's take a look at the loans delta table when we select it we can view its schema on the right hand side clicking through the tabs we can view some sample data from this table and see details about where the data files themselves are actually located and whether it is a fully managed table or not on the permissions tab we can grant access to groups or individual users let's go ahead and add the data scientists group as well as our friend omit perfect as you can see those permissions are now recorded here on the right hand side finally when you need to automate your permission management you can grant or revoke access to data objects in the lake house using standard sql grant statements like the ones seen here simply run these statements using the built-in sql query editor and databricks sql updates the permissions immediately and if you want you can also synchronize these access controls directly with azure active directory or your cloud identity service of choice jumping back to the sql endpoints tab from here we can view and monitor the health of all of the end points that are currently running or start up a new one let's go ahead and create a new endpoint now as you can see we have some really simple options here which is great sql endpoints were designed from scratch to handle low latency high concurrency bi and sql workloads to that end it offers simple t-shirt sizing for end points so that you can quickly adjust endpoint capacity to hit the right latency for your needs in addition multi-cluster load balancing allows you to automatically launch new clusters to meet demand when the load on your endpoints becomes too high and finally the databricks photon engine is a native vectorized query engine that provides faster more efficient execution of all your sql workloads for now let's use databricks sql as a powerful sql backend for tableau and we'll demo the built-in data exploration and dashboarding tools later to set up the tableau connection first we need to copy the sql endpoint credentials from the connection details tab then we'll switch over to the user settings to create a new personal access token that will be used to authenticate to tableau now we're ready to switch over to tableau desktop let's go ahead and create a new workbook and then select databricks as our data source next just paste in the credentials from the previous steps once we sign in in the upper left hand corner you'll see that tableau is now connected to our databrick sql endpoint we can search for the database we've been working with so far delta db and select the loans delta table when we click the update now button databrick sql returns query results almost instantly now that we have a direct connection to tableau we can run interactive queries and build beautiful visualizations and dashboards using databrick sql on the back end returning now to databrick sql as data admins we'd like to be able to monitor the health and performance of our endpoints to minimize downtime we can view the status and load on our endpoints by navigating to monitoring and viewing the load balancing chart as good stewards of our data we'd also like to be able to know what data sets people accessed when they accessed them and what queries they ran for purposes of auditability lineage and data governance we can view this information by clicking the query history tab from here we can view a central log of all of the time stamped queries that users have run on a particular sql endpoint we can click on any individual query to get a general overview with some summary statistics or we can choose the execution details tab for even more granular metrics if we want to dive down even deeper we can view the execution summary the information shown here allows you to identify bottlenecks for long-running queries understand the physical plan or view detailed execution information for debugging purposes now that we've touched on the admin experience let's briefly switch gears and explore databricks sql from the data analyst's perspective if you want a quick way to get started the starter databrick sql endpoint comes pre-loaded with sample data sets queries and dashboards you can copy these as templates or just play around with them to discover features and get comfortable with the built-in sql id the new sql query editor allows you to run queries and tabs that persist even if you close your browser and return to databricks sql later on from here you can quickly convert the results of your queries into visualizations and build those into dashboards finally the past executions tab offers a quick look at your personal query history so that you can easily go back and review interesting queries thanks for watching this demo visit databricks.com try to get started with databricks sql today