Modules 4-5 updated to Bravo-Lock

This commit is contained in:
Buck Woody 2020-03-05 08:10:16 -05:00
Родитель 3088b2e9e3
Коммит 9a3671281e
2 изменённых файлов: 6 добавлений и 12 удалений

Просмотреть файл

@ -27,7 +27,7 @@ Each of these functions are available separately based on the requirements of yo
<h3><a name="2-1">4.1 Data Virtualization</a></h3>
In both a Stand-Alone Instance configuration and within the BDC (BDC) configuration, you can use a series of *Connectors* to query data using the <a href="https://docs.microsoft.com/en-us/sql/relational-databases/polybase/polybase-guide?view=sql-server-ver15" target="_blank"><i>PolyBase</i></a> feature. PolyBase enables your SQL Server instance to process Transact-SQL queries that read data from external data sources. Starting in SQL Server 2019, you can access external data in Hadoop, Azure Blob Storage and also external data in SQL Server, Oracle, Teradata, and MongoDB - as well as Generic ODBC sources. PolyBase pushes as much of the query as possible to the source system, which optimizes the query.
In both a Stand-Alone Instance configuration and within the BDC (BDC) configuration, you can use a series of *Connectors* to query data using the <a href="https://docs.microsoft.com/en-us/sql/relational-databases/polybase/polybase-guide?view=sql-server-ver15" target="_blank"><i>PolyBase</i></a> feature. PolyBase enables your SQL Server instance to process Transact-SQL queries that read data from external data sources. Starting in SQL Server 2019, you can access external data in Hadoop, Azure Blob Storage and also external data in SQL Server, Oracle, Teradata, and MongoDB. PolyBase pushes as much of the query predicate as possible to the source system, which optimizes the query.
<br>
<img style="height: 300; box-shadow: 0 4px 8px 0 rgba(0, 0, 0, 0.2), 0 6px 20px 0 rgba(0, 0, 0, 0.19);" src="https://github.com/microsoft/sqlworkshops/blob/master/sqlserver2019bigdataclusters/graphics/datavirtualization1.png?raw=true">
@ -68,7 +68,7 @@ The BDC can be deployed to three environments:
- Locally for Testing (using minikube or another such environment)
- In a Cloud Service (Such as the Azure Kubernetes Service - AKS)
- On premises (using KubeADM)
- On premises (using KubeADM, KubeSpray and others)
These architectures are not mutually exclusive - you can install some components on-premises, and others as a service. Your connections can interconnect across these environments. You'll explore more about deploying a BDC in the <i>03 Planning, Installation and Configuration</i> module.
@ -100,7 +100,7 @@ The controller service provides the following core functionality:
- Expose troubleshooting tools to detect and repair unexpected issues
- Manage cluster security: ensure secure cluster endpoints, manage users and roles, configure credentials for intra-cluster communication
- Manage the workflow of upgrades so that they are implemented safely
- Manage high availability and DR for statefull services in the cluster
- Manage high availability and DR for stateful services in the cluster
You have two ways of working with the Controller service: the `azdata` utility, and the <i>Azure Data Studio</i> tool. All communication to the controller service is conducted via a REST API over HTTPS. A self-signed certificate will be automatically generated for you at bootstrap time. Authentication to the controller service endpoint is based on username and password.
@ -123,7 +123,6 @@ The Master Instance stores meta-data which is outside the scope of the meta-data
- Details of external tables that provide access to the cluster data plane.
- PolyBase external data sources and external tables defined in user databases.
These components are used in the SQL Server Master Instance:
<table style="tr:nth-child(even) {background-color: #f2f2f2;}; text-align: left; display: table; border-collapse: collapse; border-spacing: 5px; border-color: gray;">
@ -133,7 +132,6 @@ These components are used in the SQL Server Master Instance:
</table>
<h3>BDC: Compute Pool</h3>
The Compute Pool holds one or more SQL Server Pods used for distributed processing under the direction of the SQL Server Master Instance. It makes the calls out to the PolyBase connectors for a distributed Compute layer of the BDC.
@ -217,7 +215,6 @@ In this section you will review the solution tutorial similar to the one you wil
<li><a href = "https://www.dataonstorage.com/resource/video/msignite2018/brk4021-deep-dive-on-sql-server-and-big-data/" target="_blank">Session on Big Data at at Ignite</a></li>
</ul>
<p><img style="float: left; margin: 0px 15px 15px 0px;" src="https://github.com/microsoft/sqlworkshops/blob/master/graphics/geopin.png?raw=true"><b >Next Steps</b></p>
Next, Continue to <a href="https://github.com/microsoft/sqlworkshops/blob/master/k8stobdc/KubernetesToBDC/05-datascience.md" target="_blank"><i> Module 5 - Data Science</i></a>.

Просмотреть файл

@ -8,15 +8,15 @@
<h2><img style="float: left; margin: 0px 15px 15px 0px;" src="https://github.com/microsoft/sqlworkshops/blob/master/graphics/textbubble.png?raw=true"> 05 - Using the SQL Server big data cluster on Kubernetes for Data Science </h2>
In this workshop you have covered <TODO: Explain where the student is at the moment>. The end of this Module contains several helpful references you can use in these exercises and in production.
In this workshop you have covered concepts and processes for using Kubernetes, and how to set up a SQL Server big data cluster on a Kubernetes Cluster. The end of this Module contains several helpful references you can use in these exercises and in production.
This module covers Container technologies and how they are different than Virtual Machines. You'll learn about the need for container orchestration using Kubernetes.
This module covers a complete workflow for this environment, focusing on a Data Science application.
<p style="border-bottom: 1px solid lightgrey;"></p>
<h2><img style="float: left; margin: 0px 15px 15px 0px;" src="https://github.com/microsoft/sqlworkshops/blob/master/graphics/pencil2.png?raw=true"><a name="4-0">4.0 End-To-End Solution with BDC</a></h2>
Recall from <i>The Big Data Landscape</i> module that you learned about the Wide World Importers company. <a href="https://azure-scenarios-experience.azurewebsites.net/big-data.html" target="_blank">Wide World Importers </a> (WWI) is a traditional brick and mortar business with a long track record of success, generating profits through strong retail store sales of their unique offering of affordable products from around the world. They have a traditional N-tier application that uses a front-end (mobile, web and installed) that interacts with a scale-out middle-tier software product, which in turn stores data in a large SQL Server database that has been scaled-up to meet demand.
<a href="https://azure-scenarios-experience.azurewebsites.net/big-data.html" target="_blank">Wide World Importers </a> (WWI) is a traditional brick and mortar business with a long track record of success, generating profits through strong retail store sales of their unique offering of affordable products from around the world. They have a traditional N-tier application that uses a front-end (mobile, web and installed) that interacts with a scale-out middle-tier software product, which in turn stores data in a large SQL Server database that has been scaled-up to meet demand.
<br>
<img style="height: 150; box-shadow: 0 4px 8px 0 rgba(0, 0, 0, 0.2), 0 6px 20px 0 rgba(0, 0, 0, 0.19);" src="https://github.com/microsoft/sqlworkshops/blob/master/graphics/WWI-002.png?raw=true">
@ -31,11 +31,8 @@ WWI has now added web and mobile commerce to their platform, which has generated
This presented the following four challenges - the IT team at WWI needs to:
- Scale data systems to reach more consumers
- Unlock business insights from multiple sources of structured and unstructured data
- Apply deep analytics with high-performance responses
- Enable AI into apps to actively engage with customers
<br>