Metadata-Version: 2.2
Name: arktci-airflow-provider-filemaker
Version: 2.4.0
Summary: Apache Airflow provider for FileMaker Cloud
Author-email: "Josh Lipton @ ArkTCI" <josh@arktci.com>
License: Apache License 2.0
Project-URL: Homepage, https://github.com/ArkTCI/airflow-providers
Classifier: Programming Language :: Python :: 3
Classifier: License :: OSI Approved :: Apache Software License
Classifier: Operating System :: OS Independent
Requires-Python: >=3.8
Description-Content-Type: text/markdown
Requires-Dist: apache-airflow>=2.0.0
Requires-Dist: boto3>=1.16.0
Requires-Dist: requests>=2.25.0
Requires-Dist: pandas>=1.0.0
Requires-Dist: pycognito>=2023.01.0

# FileMaker Cloud Provider for Apache Airflow

This is a custom provider package for Apache Airflow that enables integration with FileMaker Cloud's OData API.

## Features

- **FileMakerHook**: Handles authentication with FileMaker Cloud through AWS Cognito and provides methods to interact with the OData API.
- **FileMakerQueryOperator**: Executes OData queries against FileMaker Cloud.
- **FileMakerExtractOperator**: Extracts data from FileMaker Cloud and saves it in various formats.
- **FileMakerSchemaOperator**: Retrieves and parses the FileMaker Cloud OData metadata schema.
- **FileMakerDataSensor**: Sensor that monitors FileMaker tables for specific conditions.
- **FileMakerChangeSensor**: Sensor that detects changes in FileMaker tables since a timestamp.
- **FileMakerCustomSensor**: Customizable sensor that allows for complex monitoring logic.

## Installation

### Installation from PyPI (Recommended)

```bash
pip install arktci-airflow-provider-filemaker
```

### Provider Structure

This provider follows the official Apache Airflow provider structure:

```
providers/filemaker/
├── pyproject.toml
├── provider.yaml
├── setup.py
├── README.md
├── src/
│   └── airflow/
│       └── providers/filemaker/
│           ├── __init__.py
│           ├── hooks/
│           │   ├── __init__.py
│           │   └── filemaker.py
│           ├── operators/
│           │   ├── __init__.py
│           │   └── filemaker.py
│           ├── sensors/
│           │   ├── __init__.py
│           │   └── filemaker.py
│           └── auth/
│               ├── __init__.py
│               └── cognitoauth.py
└── tests/
    ├── unit/
    │   └── filemaker/
    │       ├── hooks/
    │       │   └── test_filemaker.py
    │       ├── operators/
    │       │   └── test_filemaker.py
    │       ├── sensors/
    │       │   └── test_filemaker.py
    │       └── auth/
    │           └── test_cognitoauth.py
    ├── integration/
    │   └── filemaker/
    │       └── test_integration_filemaker.py
    └── system/
        └── filemaker/
            └── example_filemaker.py
```

### Manual Installation

1. Copy the `providers/filemaker` directory to your Airflow project's `providers` directory.

2. Install the required dependencies:
   ```
   pip install -r requirements.txt
   ```

3. Install the package:
   ```
   pip install -e .
   ```

4. Create a FileMaker connection in Airflow:
   - Connection ID: `filemaker_default` (or any ID you prefer)
   - Connection Type: `filemaker`
   - Host: Your FileMaker Cloud host (e.g., `my-fmcloud.filemaker-cloud.com`)
   - Schema: Your FileMaker database name
   - Login: Your FileMaker Cloud username (Claris ID)
   - Password: Your FileMaker Cloud password
   - Extra: JSON containing Cognito details (if not using auto-discovery):
     ```json
     {
       "user_pool_id": "your-cognito-user-pool-id",
       "client_id": "your-cognito-client-id",
       "region": "your-aws-region"
     }
     ```

### Connection Testing

The FileMaker provider supports connection testing in the Airflow UI. To test your connection:

1. After configuring your connection as described above, click the **Test** button.
2. The provider will attempt to authenticate with your FileMaker Cloud instance using the provided credentials.
3. You'll receive feedback indicating whether the connection was successful or if there were any issues.

The test verifies:
- All required connection parameters are provided (host, database, username, password)
- Authentication with FileMaker Cloud is successful
- A valid token can be retrieved

Common error messages and their solutions:
- **Missing FileMaker host/database/username/password**: Ensure all required fields are filled in.
- **Failed to retrieve authentication token**: Verify your credentials are correct.
- **Connection failed**: Check network connectivity to your FileMaker server.

For more detailed information about setting up connections, refer to the [connections documentation](./docs/connections.md).

## Development

### Setting Up Development Environment

1. Clone the repository:
   ```bash
   git clone https://github.com/your-org/airflow-dev.git
   cd airflow-dev/providers/filemaker
   ```

2. Install development requirements:
   ```bash
   pip install -r requirements-dev.txt
   ```

3. Set up your credentials for integration tests:
   - Copy `.env.template` to `.env`
   - Fill in your FileMaker Cloud credentials

```bash
cp .env.template .env
# Edit .env with your credentials
```

### Running Tests

#### Unit Tests

Unit tests can be run without FileMaker Cloud credentials:

```bash
pytest tests/unit
```

#### Integration Tests

Integration tests require valid FileMaker Cloud credentials.

**Important Note on Authentication:** The integration tests use AWS Cognito authentication with the `/fmi/odata/login/info` endpoint. If your FileMaker Cloud instance uses a different authentication method or if this endpoint is not accessible, the tests will fail with a 400 Bad Request error. You may need to modify the authentication code to match your FileMaker Cloud version's requirements.

To run integration tests:

1. Ensure your `.env` file is properly configured with:
   - FILEMAKER_HOST (must include https:// protocol)
   - FILEMAKER_DATABASE
   - FILEMAKER_USERNAME
   - FILEMAKER_PASSWORD

2. Run the tests:
```bash
./run_tests.sh
```

Or run specific tests:
```bash
python -m pytest tests/integration -v
```

### Continuous Integration

For CI/CD with GitHub Actions, add the following secrets to your repository:

- FILEMAKER_HOST
- FILEMAKER_DATABASE
- FILEMAKER_USERNAME
- FILEMAKER_PASSWORD

This will allow integration tests to run during the release workflow.

## Troubleshooting

### Integration Test Authentication Failures

If you're seeing authentication errors like `400 Client Error: Bad Request for url: https://your-host/fmi/odata/login/info`, it may indicate:

1. Your FileMaker Cloud instance is using a different authentication endpoint
2. The FileMaker Cloud version has updated its API format
3. Your credentials don't have the necessary permissions

Check your FileMaker Cloud documentation for the correct API endpoint format and authentication requirements.

## License

This provider is licensed under the same license as Apache Airflow.

**Important Note**: If you use placeholder values like "your-filemaker-host.com" in your .env file, the integration tests will attempt to run but will fail with connection errors. This is expected behavior since the placeholders are not valid hosts. You need real FileMaker Cloud credentials for the integration tests to pass.

To run only specific tests, you can use pytest directly:

```bash
# Run only unit tests
python -m pytest tests/unit -v

# Run only integration tests
python -m pytest tests/integration -v

# Run a specific test file
python -m pytest tests/unit/filemaker/hooks/test_filemaker.py -v
```

When running integration tests in a CI/CD environment, you'll need to securely provide credentials:

1. In your GitHub repository, go to Settings > Secrets and add the following secrets:
   - `FILEMAKER_HOST` (include the https:// protocol)
   - `FILEMAKER_DATABASE`
   - `FILEMAKER_USERNAME`
   - `FILEMAKER_PASSWORD`

2. These secrets will be automatically used by the GitHub Actions workflow when running integration tests.

## Authentication Information

### OData Authentication Details

This provider implements authentication using AWS Cognito with the following fixed credentials (same as the JavaScript implementation):

- **User Pool ID**: `us-west-2_NqkuZcXQY`
- **Client ID**: `4l9rvl4mv5es1eep1qe97cautn`

The authentication header format used is:
```
Authorization: FMID <token>
```

This matches the implementation in JavaScript examples like:

```javascript
// JavaScript example
const response = await fetch(baseUrl, {
  headers: {
    'Authorization': `FMID ${fmidToken}`,
    'Accept': 'application/json'
  }
});
```

This authentication implementation:
1. Uses a standard AWS Cognito authentication flow (`USER_PASSWORD_AUTH`)
2. Retrieves an ID token for your FileMaker credentials
3. Uses the ID token with the "FMID" prefix in the Authorization header

### Authentication Troubleshooting

If you're experiencing authentication issues:

1. Ensure your FileMaker host URL includes the protocol (`https://`)
2. Verify your credentials are correct
3. Check network logs to confirm the token is being sent correctly with the `FMID` prefix

The integration tests have been enhanced to provide detailed debugging information to help diagnose authentication issues.

## Integration Tests

The integration tests for this provider require valid FileMaker Cloud credentials. The tests will authenticate with FileMaker Cloud and then run a series of tests to validate the functionality of the provider.

To run the integration tests:

1. Ensure your `.env` file is properly configured with:
   - FILEMAKER_HOST (must include https:// protocol)
   - FILEMAKER_DATABASE
   - FILEMAKER_USERNAME
   - FILEMAKER_PASSWORD

2. Run the tests:
```bash
./run_tests.sh
```

Or run specific tests:
```bash
python -m pytest tests/integration -v
```

When running integration tests in a CI/CD environment, you'll need to securely provide credentials:

1. In your GitHub repository, go to Settings > Secrets and add the following secrets:
   - `FILEMAKER_HOST` (include the https:// protocol)
   - `FILEMAKER_DATABASE`
   - `FILEMAKER_USERNAME`
   - `FILEMAKER_PASSWORD`

2. These secrets will be automatically used by the GitHub Actions workflow when running integration tests.

## Operators

### FileMakerQueryOperator

This operator executes a query against a FileMaker database and returns the results.

```python
from airflow.providers.filemaker.operators.filemaker import FileMakerQueryOperator

query_task = FileMakerQueryOperator(
    task_id="query_filemaker",
    filemaker_conn_id="filemaker_default",
    database="students",
    layout="students",
    query={"Grade": "A"},
)
```

### FileMakerExtractOperator

This operator extracts data from a FileMaker database and saves it to a file.

```python
from airflow.providers.filemaker.operators.filemaker import FileMakerExtractOperator

extract_task = FileMakerExtractOperator(
    task_id="extract_filemaker",
    filemaker_conn_id="filemaker_default",
    database="students",
    layout="students",
    query={"Grade": "A"},
    destination_path="/tmp/students.json",
)
```

### FileMakerSchemaOperator

This operator retrieves the schema for a FileMaker layout.

```python
from airflow.providers.filemaker.operators.filemaker import FileMakerSchemaOperator

schema_task = FileMakerSchemaOperator(
    task_id="get_schema",
    filemaker_conn_id="filemaker_default",
    database="students",
    layout="students",
)
```

### FileMakerCreateRecordOperator

This operator creates a new record in a FileMaker database.

```python
from airflow.providers.filemaker.operators.filemaker import FileMakerCreateRecordOperator

create_task = FileMakerCreateRecordOperator(
    task_id="create_record",
    filemaker_conn_id="filemaker_default",
    database="students",
    layout="students",
    record_data={
        "FirstName": "John",
        "LastName": "Doe",
        "Email": "john.doe@example.com",
        "Grade": "A",
    },
)
```

### FileMakerUpdateRecordOperator

This operator updates an existing record in a FileMaker database.

```python
from airflow.providers.filemaker.operators.filemaker import FileMakerUpdateRecordOperator

update_task = FileMakerUpdateRecordOperator(
    task_id="update_record",
    filemaker_conn_id="filemaker_default",
    database="students",
    layout="students",
    record_id="123",  # Can also use XCom: "{{ ti.xcom_pull(task_ids='create_record')['recordId'] }}"
    record_data={
        "Grade": "A+",
        "Notes": "Updated via Airflow",
    },
)
```

### FileMakerDeleteRecordOperator

This operator deletes a record from a FileMaker database.

```python
from airflow.providers.filemaker.operators.filemaker import FileMakerDeleteRecordOperator

delete_task = FileMakerDeleteRecordOperator(
    task_id="delete_record",
    filemaker_conn_id="filemaker_default",
    database="students",
    layout="students",
    record_id="123",  # Can also use XCom: "{{ ti.xcom_pull(task_ids='create_record')['recordId'] }}"
)
```

### FileMakerBulkCreateOperator

This operator creates multiple records in a FileMaker database in a single request.

```python
from airflow.providers.filemaker.operators.filemaker import FileMakerBulkCreateOperator

bulk_create_task = FileMakerBulkCreateOperator(
    task_id="bulk_create",
    filemaker_conn_id="filemaker_default",
    database="students",
    layout="students",
    records_data=[
        {
            "FirstName": "Jane",
            "LastName": "Smith",
            "Email": "jane.smith@example.com",
            "Grade": "B+",
        },
        {
            "FirstName": "Bob",
            "LastName": "Johnson",
            "Email": "bob.johnson@example.com",
            "Grade": "C",
        },
    ],
)
```

### FileMakerExecuteFunctionOperator

This operator executes a FileMaker script/function.

```python
from airflow.providers.filemaker.operators.filemaker import FileMakerExecuteFunctionOperator

execute_script_task = FileMakerExecuteFunctionOperator(
    task_id="execute_script",
    filemaker_conn_id="filemaker_default",
    database="students",
    layout="students",
    script_name="UpdateGrades",
    script_params={
        "gradeThreshold": "C",
        "newGrade": "C+",
    },
)
```

### FileMakerToS3Operator

This operator extracts data from a FileMaker database and uploads it to an S3 bucket.

```python
from airflow.providers.filemaker.operators.filemaker import FileMakerToS3Operator

to_s3_task = FileMakerToS3Operator(
    task_id="filemaker_to_s3",
    filemaker_conn_id="filemaker_default",
    aws_conn_id="aws_default",
    database="students",
    layout="students",
    query={"Grade": "A"},
    s3_bucket="my-bucket",
    s3_key="data/students.json",
)
```

## Example DAGs

The provider package includes example DAGs that demonstrate the usage of the FileMaker operators:

1. `example_filemaker_query.py` - Shows how to query data from FileMaker
2. `example_filemaker_to_s3.py` - Shows how to extract data from FileMaker and upload it to S3
3. `example_filemaker_record_management.py` - Shows how to create, update, and delete records in FileMaker 
