diff --git a/.github/workflows/release.yml b/.github/workflows/release.yml index c6df5454..aaa5d8f0 100644 --- a/.github/workflows/release.yml +++ b/.github/workflows/release.yml @@ -27,7 +27,7 @@ jobs: mkdir alphasql_darwin_amd64 mv ./bazel-bin/alphasql/dag ./alphasql_darwin_amd64/dag mv ./bazel-bin/alphasql/pipeline_type_checker ./alphasql_darwin_amd64/pipeline_type_checker - tar -zcvf bin/alphasql_darwin_amd64.tar.gz alphasql_darwin_amd64 + tar -zcvf alphasql_darwin_amd64.tar.gz alphasql_darwin_amd64 - name: Get Release id: get uses: bruceadams/get-release@v1.2.0 @@ -37,7 +37,7 @@ jobs: GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} with: upload_url: ${{ steps.get.outputs.upload_url }} # This pulls from the CREATE RELEASE step above, referencing it's ID to get its outputs object, which include a `upload_url`. See this blog post for more info: https://jasonet.co/posts/new-features-of-github-actions/#passing-data-to-future-steps - asset_path: ./bin/alphasql_darwin_amd64.tar.gz + asset_path: alphasql_darwin_amd64.tar.gz asset_name: alphasql_darwin_amd64.tar.gz asset_content_type: application/tar+gzip @@ -72,7 +72,7 @@ jobs: mkdir alphasql_linux_x86_64 mv ./bazel-bin/alphasql/dag ./alphasql_linux_x86_64/dag mv ./bazel-bin/alphasql/pipeline_type_checker ./alphasql_linux_x86_64/pipeline_type_checker - tar -zcvf bin/alphasql_linux_x86_64.tar.gz alphasql_linux_x86_64 + tar -zcvf alphasql_linux_x86_64.tar.gz alphasql_linux_x86_64 - name: Get Release id: get uses: bruceadams/get-release@v1.2.0 @@ -83,7 +83,7 @@ jobs: GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }} with: upload_url: ${{ steps.get.outputs.upload_url }} # This pulls from the CREATE RELEASE step above, referencing it's ID to get its outputs object, which include a `upload_url`. See this blog post for more info: https://jasonet.co/posts/new-features-of-github-actions/#passing-data-to-future-steps - asset_path: ./bin/alphasql_linux_x86_64.tar.gz + asset_path: alphasql_linux_x86_64.tar.gz asset_name: alphasql_linux_x86_64.tar.gz asset_content_type: application/tar+gzip diff --git a/README.md b/README.md index 7d6544df..8c8cbdbb 100644 --- a/README.md +++ b/README.md @@ -61,10 +61,10 @@ wget -P $temp https://github.com/Matts966/alphasql/releases/latest/download/alph ```bash # To extract DAG from your SQL set -$ ./bin/osx/dag --output_path ./samples/sample1/dag.dot ./samples/sample1/ +$ dag --output_path ./samples/sample1/dag.dot ./samples/sample1/ # Or you can check the output in stdout by -$ ./bin/osx/dag [paths] +$ dag [paths] # with graphviz $ dot -Tpng samples/sample1/dag.dot -o samples/sample1/dag.png @@ -109,7 +109,7 @@ Note that you should run type_checker in the same path as in extracting DAG. ```bash # to check type and schema of SQL set -$ ./bin/osx/pipeline_type_checker ./samples/sample1.dot +$ pipeline_type_checker ./samples/sample1.dot Analyzing "./samples/sample1/create_datawarehouse3.sql" DDL analyzed, adding table to catalog... SUCCESS: analysis finished! @@ -169,13 +169,13 @@ You can specify external schemata (not created by queries in SQL set) by passing ```bash # with external schema -$ ./bin/osx/pipeline_type_checker --json_schema_path ./samples/sample-schema.json ./samples/sample1/dag.dot +$ pipeline_type_checker --json_schema_path ./samples/sample-schema.json ./samples/sample1/dag.dot ``` You can extract required external tables by ```bash -$ ./bin/osx/dag --external_required_tables_output_path ./required_tables.txt {./path/to/sqls} +$ dag --external_required_tables_output_path ./required_tables.txt {./path/to/sqls} # and get schemata using bq command $ cat ./required_tables.txt | while read line do diff --git a/docker/linux-copy-bin.sh b/docker/linux-copy-bin.sh deleted file mode 100755 index a352129a..00000000 --- a/docker/linux-copy-bin.sh +++ /dev/null @@ -1,6 +0,0 @@ -#!/bin/bash -set -ex -id=$(docker create matts966/alphasql /bin/bash) -docker cp $id:/usr/bin/dag ./bin/linux/dag -docker cp $id:/usr/bin/pipeline_type_checker ./bin/linux/pipeline_type_checker -docker rm -v $id