To Developers,
I am doing benchmarks for Azure Data Lake and I am seeing about ~7.5 MB/S for a read of an ADL Store and a write to a VHD all in the same region. This is the case for PowerShell and C# with the code taken from the following examples:
PowerShell Code is from https://azure.microsoft.com/en-us/documentation/articles/data-lake-store-get-started-powershell/ C# Code is from https://azure.microsoft.com/en-us/documentation/articles/data-lake-store-get-started-net-sdk/
Are the above code samples acceptable for a benchmark test or will a new SDK be delivered that will enhance the throughput? Also, are there expected throughput numbers when ADL Store becomes generally available?
Thanks, Marc
I started to write an Azure Data Lake Storage Throughput Analyzer and put the first code bits on GitHub.
You should run that tool on an Azure VM to not measure you internet connection.
Please feel free to add you thoughts and code contributions to my GitHub repo as well.
I hope this helps.