SheildPy

Protect sensitive data with SheildPy, a Python package offering encryption, anonymization, and compliance tools for data scientists.


License
MIT
Install
pip install SheildPy==0.0.8

Documentation

secure-learn: Secure Data Privacy Framework for Python Data Scientists

image

image

secure-learn is an all-in-one Python package designed to address data privacy and security concerns for data scientists. Developed by Deependra Verma, secure-learn offers robust encryption, anonymization, and access control tools, ensuring the confidentiality and integrity of sensitive data.

Introduction

secure-learn: Your all-in-one Python package for robust data privacy and security. Encrypt, anonymize, and control access to sensitive data effortlessly.

Features

secure-learn provides the following key methods:

  • encrypt_data(data): Encrypts sensitive data to ensure confidentiality.
  • decrypt_data(encrypted_data): Decrypts encrypted data to its original form.
  • anonymize_data(data, columns_to_anonymize): Anonymizes specific columns in a DataFrame.
  • add_role(role_name, permissions): Adds a new role with associated permissions to the access control system.
  • check_permission(role_name, permission): Checks if a role has the specified permission.

Installation

To install secure-learn, simply run:

Alternatively, you can clone the GitHub repository:

Dependencies

secure-learn relies on the following dependencies:

  • pandas>=1.0.0
  • faker>=8.0.0
  • cryptography>=3.0

Usage

Import the package:

Initialize PrivacyPy with encryption key:

Anonymize sensitive columns (Name, Email):

Encrypt entire DataFrame:

Users Benefit

secure-learn empowers data scientists with the following benefits:

  • Data Confidentiality: Encrypt sensitive data to prevent unauthorized access.
  • Anonymization: Anonymize personally identifiable information for privacy protection.
  • Access Control: Control data access based on user roles and permissions.
  • Compliance: Ensure compliance with data protection regulations (e.g., GDPR, HIPAA).

Use Cases

secure-learn can be used in various data science scenarios, including:

  • Healthcare data analysis
  • Financial data processing
  • User authentication systems
  • Research collaborations with external parties

Invitation for Contribution

Contributions to secure-learn are welcome! To contribute, follow these steps:

  1. Fork the repository on GitHub.
  2. Clone the forked repository to your local machine.
  3. Create a new branch for your changes.
  4. Make your modifications and improvements.
  5. Test your changes to ensure they work as expected.
  6. Commit your changes and push them to your forked repository.
  7. Submit a pull request to the original repository.

We welcome contributions from the community! Whether it's fixing bugs, adding new features, or improving documentation, your contributions help make XplainML better for everyone. Check out our Contributing Guidelines to get started.

License

secure-learn is licensed under the MIT License. See the LICENSE file for details.

About the Author

Email | LinkedIn | GitHub | Portfolio