* Testing - Clean up the Argo controller that was used to build images
* Try to not install Argo in the first place
* Added the test-runner service account
* Cleanup
* Changing the install-argo.sh script instead
* Allows toggling between one-off and recurring runs in the new run page
* Clean up and adds tests
* Fix integration test - account for extra field in form
* Cleanup and PR comments
* Clear default exp table on delete and create default exp on run create
if no default exists
With this change, if the delete experiment API is called on the default
experiment, then the ID will also be removed from the default_experiments
table.
Additionally, if the default experiment doesn't exist and a new run is
created without an experiment, a new default experiment will be created,
and the run will be placed within this experiment.
* Adds integration test for creating a run without an experiment
* Fixes failure to close database connection and adds tests for recreating and deleting default experiment
* Rename function
* Revert some row.Close() calls
* Testing/Sample - Made checking confusion matrix data more robust
The sample tests no longer depend on particular file names inside the archive. Now they only depend on the artifact name.
* Fixed json loading on python 3.5
`json.load` only supports reading from binary files in python 3.6+. https://docs.python.org/3/library/json.html#json.load
* Adds 'Create run' button to experiment list / all runs page
* Add run without experiment and filtering to FE integration test
* Update snapshots
* Add refresh and wait to integration test
* Adjust
* Adjust
* Don't exit integration test early if npm test fails
* PR comments
* TEMP - take screenshots to debug integration test
* Store screenshots
* Remove create run without experiment integration test for now as it fails due to the default experiment being deleted at the end of the API initialization and integration test suites
* WIP - Create default experiment upon API server initialization
* Default experiment initialization caused crashes if API server pod was restarted without clearing DB
* Adding new table to store default experiment ID
* Add default experiment type model definition
* Minor fixes, everything seems to work now
* Clean up. Renamed to default_experiment_store
* Adds tests for the default_experiment_store
* Add integration test for verifying initial cluster state. Currently only covers existence of default experiment
* Don't run initialization tests except as integration tests
* Fixes comments
* PR comments and cleanup
* Extract code in resource_manager to helper func
* add type checking sample to sample tests
* Dadd the test script exit code to the sample test result; update the check_notebook_result script to not validate the pipeline runs when experiment arg is not provided
* fix typo
* Fixed Kubeflow sample test
* Fixed the artifact-finding logic in `get_artifact_in_minio`.
It was just taking the first artifact before.
Now it properly searches the artifact by name.
* Update swagger definitions
* WIP - Adds ability to terminate runs to frontend
* Update snapshots
* Adds tests. Also changes warning message color to orange rather than red
* Remove refresh button from run details page
* Elaborate terminate confirmation message
* Minor fixes
* Remove references to refresh button from integration tests
* add a While in the ops group
* deepcopy the while conditions when entering and exiting
* add while condition resolution in the compiler
* define graph component decorator
* remove while loop related codes
* fixes
* remove while loop related code
* fix bugs
* generate a unique ops group name and being able to retrieve by name
* resolve the opsgroups inputs and dependencies based on the pipelineparam in the condition
* add a recursive ops_groups
* fix bugs of the recursive opsgroup template name
* resolve the recursive template name and arguments
* add validity checks
* add more comments
* add usage comment in graph_component
* add a sample
* add unit test for the graph opsgraph
* refactor the opsgroup
* add unit test for the graph_component decorator
* exposing graph_component decorator
* add recursive compiler unit tests
* add the sample test
* fix the bug of opsgroup name
adjust the graph_component usage example
fix index bugs
use with statement in the graph_component instead of directly calling
the enter/exit functions
* add a todo to combine the graph_component and component decorators
* fix some merging bug
* fix typo
* add more comments in the sample
* update comments
* dsl generate zip file
* minor fix
* fix zip read in the unit test
* update sample tests
* dsl compiler generates pipeline based on the input name suffix
* add unit tests for different output format
* update the sdk client to support tar zip and yaml
* fix typo
* fix file write
* add postsubmit script and yaml
* remove old sample tests component file
* extract deploy-pipeline.sh, deploy-kubeflow.sh and test-prep.sh from presubmit and postsubmit scripts
* component build support for both python2 and python3
* add sample test
* remove the annotations for python2 component build
* add pathlib for python2 component build
* fix component build unit test
* fix bug in the dockerfile generator
* remove exist_ok in path.mkdir to make python2 compatible
* adjust unit test
* remove pathlib dependency for python2 component build
* remove the pathlib codes in python3 component build, but use python2 code instead; add a todo to create a new sample
* merge build image to test suit
* update image
* Update presubmit-tests-with-pipeline-deployment.sh
* add permission to access to gcr
* add service account
* test
* fix
* not exit
* speed boost
* Uses 'Create' for all actions that lead to creation flow, or result in a static object (experiment). 'Start' is used solely for initiating runs
* Update integration test
* Update sample notebook to clean up deployed models.
Update SDK client to return correct links in local Jupyter with user's own proxy connection.
* Fix sample tests.
* add another sample test to test the current sample codes instead of using newly built component images
* rename sample test yamls
* use the v2 name
* bash bug
* tf-training bug fix
* output argo log in case of exceptions for tf-training sample
* disable gpu
* try go version in travis
* add back old travis tests with backend tests
* remove backend unit tests prow config
* remove unit_test_gke
* test backend/src directory
* update comment to call out unit test