Developer Instructions ====================== Guidance for developers. Pre-Commit Hooks ---------------- We use the excellent `pre-commit `_ to run several hooks on all changes before commits. ``pre-commit`` is included in the ``dev`` extra installs. You'll have to run ``pre-commit install`` once per environment before committing changes. The reason behind running black, isort, and others as a pre-commit hook is to let a machine make style decisions, based on the collective wisdom of the Python community. Generating Documentation ------------------------ You will need to ``pip install`` the ``dev`` requirements:: pip install -e .[dev] Then from the root of the repo you can type:: make sphinx This will automatically regenerate the api documentation using ``sphinx-apidoc``. The rendered documentation will be stored in the ``/docs/build`` directory. The generated documentation is served from the ``gh-pages`` branch. Make sure that the branch is clean and then to push to gh-pages you can type:: make ghpages Note about documentation: The `Numpy and Google style docstrings `_ are activated by default. Just make sure Sphinx 1.3 or above is installed. Run unit tests -------------- Run ``python -m pytest`` to run all unittests defined in the subfolder ``tests`` with the help of `py.test `_ and `pytest-runner `_. Snowflake testing ----------------- Testing the Snowflake compare requires the use of a Snowflake cluster, as Snowflake does not support local running. This means that Snowflake tests do not get run in CICD, and changes to the Snowflake Compare must be validated by the process of running these tests locally. Note that you must have the following environment variables set in order to instantiate a Snowflake Connection (for testing purposes): - "SF_ACCOUNT": with your SF account - "SF_UID": with your SF username - "SF_PWD": with your SF password - "SF_WAREHOUSE": with your desired SF warehouse - "SF_DATABASE": with a valid database with which you have access - "SF_SCHEMA": with a valid schema belonging to the provided database Once these are set, you are free to run the suite of Snowflake tests. Management of Requirements -------------------------- Requirements of the project should be added to ``pyproject.toml``. Optional requirements used only for testing, documentation, or code quality are added to ``pyproject.toml`` in the ``project.optional-dependencies`` section. edgetest -------- edgetest is a utility to help keep requirements up to date and ensure a subset of testing requirements still work. More on edgetest `here `_. The ``pyproject.toml`` has configuration details on how to run edgetest. This process can be automated via GitHub Actions. (A future addition, which will come soon). In order to execute edgetest locally you can run the following after install ``edgetest``: .. code-block:: bash edgetest -c pyproject.toml --export This should return output like the following and also updating ``pyproject.toml``: .. code-block:: bash ============= =============== =================== ================= Environment Passing tests Upgraded packages Package version ============= =============== =================== ================= core True boto3 1.21.7 core True pandas 1.3.5 core True PyYAML 6.0 ============= =============== =================== ================= Release Guide ------------- For ``datacompy`` we want to use a simple workflow branching style and follow `Semantic Versioning `_ for each release. ``develop`` is the default branch where most people will work with day to day. All features must be squash merged into this branch. The reason we squash merge is to prevent the develop branch from being polluted with endless commit messages when people are developing. Squashing collapses all the commits into one single new commit. It will also make it much easier to back out changes if something breaks. ``main`` is where official releases will go. Each release on ``main`` should be tagged properly to denote a "version" that will have the corresponding artifact on pypi for users to ``pip install``. ``gh-pages`` is where official documentation will go. After each release you should build the docs and push the HTML to the pages branch. When first setting up the repo you want to make sure your gh-pages is a orphaned branch since it is disconnected and independent from the code: ``git checkout --orphan gh-pages``. The repo has a ``Makefile`` in the root folder which has helper commands such as ``make sphinx``, and ``make ghpages`` to help streamline building and pushing docs once they are setup right. Generating distribution archives (PyPI) --------------------------------------- After each release the package will need to be uploaded to PyPi. The instructions below are taken from `packaging.python.org `_ Update / Install ``build``, ``wheel``, and ``twine``:: pip install --upgrade build wheel twine Generate distributions:: python -m build Under the ``dist`` folder you should have something as follows:: dist/ datacompy-0.1.0-py3-none-any.whl datacompy-0.1.0.tar.gz Finally upload to PyPi:: # test pypi twine upload --repository-url https://test.pypi.org/legacy/ dist/* # real pypi twine upload dist/*