Data Accelerator for Apache Spark simplifies onboarding to Streaming of Big Data. It offers a rich, easy to use experience to help with creation, editing and management of Spark jobs on Azure HDInsights or Databricks while enabling the full power of the Spark engine.
Перейти к файлу
Yinyu Wang f5997b2904 refactor listFiles 2022-09-13 09:15:49 +00:00
.github Pointing to each pom and package.json (#256) 2022-06-30 11:12:41 -07:00
.vscode Remove unnecessary keyvault access for ABFS file URI. (#227) 2022-05-03 15:33:09 -07:00
DataProcessing refactor listFiles 2022-09-13 09:15:49 +00:00
DeploymentCloud Update dockerfile to pull in dotnet image from mcr (#235) 2022-06-01 22:35:38 -07:00
DeploymentLocal Update dockerfile to pull in dotnet image from mcr (#235) 2022-06-01 22:35:38 -07:00
Docs Formating and other updates to docs. (#9) 2019-04-16 15:07:33 -07:00
Services Bump Microsoft.NetFramework.Analyzers from 2.6.3 to 3.3.2 in /Services (#241) 2022-07-11 08:34:34 -07:00
Tests Bump Newtonsoft.Json from 12.0.2 to 13.0.1 in /Services/JobRunner (#245) 2022-06-24 09:58:22 -07:00
Website Bump async from 2.6.3 to 2.6.4 in /Website/Website (#508) 2022-08-26 19:08:04 -07:00
.gitignore provide a scenario tester to run through actions on a host in sequence and parallelly. Enables creating simulated test loads. 2019-07-15 14:16:29 -07:00
CODE_OF_CONDUCT.md Initial Checkin 2019-04-15 23:57:37 -07:00
CONTRIBUTING.md Include status badge on Readme (#37) 2019-04-25 11:43:17 -07:00
LICENSE Initial Checkin 2019-04-15 23:57:37 -07:00
README.md Updated README (#165) 2019-11-20 10:49:10 -08:00
SECURITY.md Microsoft mandatory file (#232) 2022-05-27 14:41:44 -07:00
ThirdPartyNotices.txt Initial Checkin 2019-04-15 23:57:37 -07:00

README.md

Data Accelerator for Apache Spark

Flow Build status Gateway Build status DataProcessing Build status
Metrics Build status SimulatedData Build status Website Build status

Data Accelerator for Apache Spark democratizes streaming big data using Spark by offering several key features such as a no-code experience to set up a data pipeline as well as fast dev-test loop for creating complex logic. Our team has been using the project for two years within Microsoft for processing streamed data across many internal deployments handling data volumes at Microsoft scale. It offers an easy to use platform to learn and evaluate streaming needs and requirements. We are thrilled to share this project with the wider community as open source!

Azure Friday: We are now featured on Azure Fridays! See the video here.

Data Accelerator offers three level of experiences:

  • The first requires no code at all, using rules to create alerts on data content.
  • The second allows to quickly write a Spark SQL query with additions like LiveQuery, time windowing, in-memory accumulator and more.
  • The third enables integrating custom code written in Scala or via Azure functions.

You can get started locally for Windows, macOs and Linux following these instructions
To deploy to Azure, you can use the ARM template; see instructions deploy to Azure.

The data-accelerator repository contains everything needed to set up an end-to-end data pipeline. There are many ways you can participate in the project:

Getting Started

To unleash the full power Data Accelerator, deploy to Azure and check cloud mode tutorials.

We have also enabled a "hello world" experience that you try out locally by running docker container. When running locally there are no dependencies on Azure, however the functionality is very limited and only there to give you a very cursory overview of Data Accelerator. To run Data Accelerator locally, deploy locally and then check out the local mode tutorials.

Data Accelerator for Spark runs on the following:

  • Azure HDInsight with Spark 2.4 (2.3 also supported)
  • Azure Databricks with Spark 2.4
  • Service Fabric (v6.4.637.9590) with
    • .NET Core 2.2
    • ASP.NET
  • App Service with Node 10.6

See the wiki pages for further information on how to build, diagnose and maintain your data pipelines built using Data Accelerator for Spark.

Contributing

If you are interested in fixing issues and contributing to the code base, we would love to partner with you. Try things out, join in the design conversations and make pull requests.

Feedback

Please also see our Code of Conduct.

Security issues

Security issues and bugs should be reported privately, via email, to the Microsoft Security Response Center (MSRC) secure@microsoft.com. You should receive a response within 24 hours. If for some reason you do not, please follow up via email to ensure we received your original message. Further information, including the MSRC PGP key, can be found in the Security TechCenter.

License

This repository is licensed with the MIT license.