Closed senthilkungumaraj closed 2 years ago
lets add a comment here that this guid is for contributor role basically granting everything
Lets address this
lets add a comment here that this guid is for contributor role basically granting everything
Thanks, @sushilkm. Good catch. Added the comment to clarify the guid.
should we call the
batch.account.role.assignment.bicep
directly from themain.bicep
makingpipelineModule
a dependency here would slow down the deployment.
@sushilkm I tried adding it to batch account first. Synapse workspace and Batch Account are being created parallel. Synapse workspace's managed identity is system assigned. So for the RBAC on batch account to work, we need Synapse workspace to complete provisioning before batch account.
Maybe we can look into optimizing this further in v2.
• The bash scripts mentioned in the README are not executable -> users will need to run chmod +x on them manually Execute permission added to executable files in deploy directory • The README asks the user to wait until the Batch pool's resizing is done, but as the number of dedicated nodes is set to 0, it will never resize • Before I was able to push the Docker container image to the ACR I had to login using
az acr login -n <name of registry>
-> This is not specified in the README • configure.sh expects environment.yml to be in the current directory, but the README suggests running the script from outside the deploy folder Added the base-path to use an absolute filepath instead of relative filepath. • I get two errors when running configure.sh: - In Progress o + az batch job create --id custom-vision-model-job --pool-id data-cpu-pool 'NoneType' object has no attribute 'startswith' o + az storage blob upload-batch --destination spark-jobs --account-name synhnsj57hcp --source ../src/transforms/spark-jobs incorrect usage: source must be an existing directory I think this is also related to executing outside the directory, updated to absolute path, lets check if it is fixed now. o I have attached the full output of the command to this mail • package.sh has a similary issue as configure.sh: It expects package.py to be in the current directory but is suggested to be run from outside deploy, result in: python3: can't open file 'package.py': [Errno 2] No such file or directory Updated this file also to use absolute path instead of relative path, now users should be able to execute from anywhere even if they miss instruction. • As a last step, the README mentions to follow the steps in "Deployment Procedures" before running the pipeline. I couldn't find this section anywhere, but I do guess it goes into detail on what to upload (input data as well as config files) to the raw storage account - if not, are you planning to add that information somewhere? Updated README to remove reference to “Deployment Procedures”.