Skip to content

l-kli/cedar-local-agent

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

68 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Cedar Local Agent

This crate is experimental.

The cedar-local-agent crate provides customers a useful foundation for creating asynchronous authorizers that can handle two different operational modes:

  1. The authorizer is able to cache all of your application’s policies and entity data while evaluating a request
  2. The authorizer is unable to cache all of your application's policies and entity data while evaluating a request

The cedar-local-agent crate provides a simple::Authorizer which can be built with option (1) or (2). The simple::Authorizer is constructed using policy and entity providers. These providers can be implemented by customers.

cedar-local-agent provides sample implementations of providers that implement option (1). A file system policy set provider: file::PolicySetProvider, and an entity provider: file::EntityProvider.

For more information about the Cedar language/project, please take a look at cedarpolicy.com.

Usage

Cedar Local Agent can be used in your application via the cedar-local-agent crate.

Add cedar-local-agent as a dependency in your Cargo.toml file. For example:

[dependencies]
cedar-local-agent = "2"

Quick Start

Build a local authorizer that evaluates authorization decisions using a locally stored policy set, entity store and schema.

Policy data: tests/data/sweets.cedar

Entity data: tests/data/sweets.entities.json

Schema: tests/data/sweets.schema.cedar.json

Build a policy set:

let policy_set_provider = PolicySetProvider::new(
    policy_set_provider::ConfigBuilder::default()
        .policy_set_path("tests/data/sweets.cedar")
        .build()
        .unwrap(),
)
.unwrap();

Build an entity provider:

let entity_provider = EntityProvider::new(
    entity_provider::ConfigBuilder::default()
        .entities_path("tests/data/sweets.entities.json")
        .schema_path("tests/data/sweets.schema.cedar.json")
        .build()
        .unwrap(),
)
.unwrap();

Build the authorizer:

let authorizer: Authorizer<PolicySetProvider, EntityProvider> = Authorizer::new(
    AuthorizerConfigBuilder::default()
        .entity_provider(Arc::new(entity_provider))
        .policy_set_provider(Arc::new(policy_set_provider))
        .build()
        .unwrap(),
);

Evaluate a decision:

assert_eq!(
    authorizer
        .is_authorized(&Request::new(
            Some(format!("User::\"Cedar\"").parse().unwrap()),
            Some(format!("Action::\"read\"").parse().unwrap()),
            Some(format!("Box::\"3\"").parse().unwrap()),
            Context::empty(),
        ), &Entities::empty())
        .await
        .unwrap()
        .decision(),
    Decision::Deny
)

simple::Authorizer is_authorized API Semantics

The simple::Authorizer is_authorized API takes a Cedar request and Cedar entities within the API.

pub async fn is_authorized(
    &self,
    request: &Request,
    entities: &Entities,
) -> Result<Response, AuthorizerError> { ... }

For scenarios where the same entity identifier, EID, is passed as input and returned by an EntityProvider, the input is considered the last value. This API favors last-value-wins semantics. This behavior is subject to change pending RFC-0020.

The file::PolicySetProvider and file::EntityProvider gather data when initialized and cache it in memory. No data is read from disk during an authorization decision.

Policy and entity data can be mutated on disk after the initialization of an authorizer. To account for this, functionality is provided which will refresh policy and entity data tangential to an authorization decision. To accomplish this, a minimum of two additional threads are required, for a total of three threads.

  1. The main thread handles is_authorized calls
  2. The signaler thread notifies receivers of required updates
  3. The receiver thread listens for updates

Channels are used to communicate between the signaler thread and the receiver thread. There are two provided functions for creating signaler threads. Both return a signaler thread and a tokio::broadcast::receiver as an output.

  1. clock_ticker_task periodically wakes up and signals based on a clock duration
  2. file_inspector_task periodically wakes up and checks for differences in a file using a collision resistant hashing function (SHA256) and notifies on modifications

Warning: It is important to be careful when selecting the refresh rate of the signaler which triggers a policy refresh. We have set up a RefreshRate enum which gives several options of RefreshRates of at least 15 seconds. This is fast enough for most applications but slow enough to be very unlikely to trigger any sort of throttling or performance impact on most policy set sources.

Receivers are required to be passed to a new separate thread to listen and respond to events. The update_provider_data_task handles receiving these signals in the form of an Event. Messages are handled one message at a time. The receiver thread blocks until it has successfully or unsuccessfully updated the data for the provider.

Sample usage of updating a policy set provider's data every fifteen seconds:

let (clock_ticker_signal_thread, receiver) = clock_ticker_task(RefreshRate::FifteenSeconds);

let policy_set_provider = Arc::new(PolicySetProvider::new(
    policy_set_provider::ConfigBuilder::default()
        .policy_set_path("tests/data/sweets.cedar")
        .build()
        .unwrap(),
)
.unwrap());

let update_provider_thread = update_provider_data_task(policy_set_provider.clone(), receiver);

Note: these background threads must remain in scope for the life of your application. If there is an issue updating in a background thread it will produce an error!() message but will not cause the application to crash.

This means that if for any reason, the agent cannot update its policy set due to an error, the agent will continue running with the stale policy set and will log an error. Since the agent will log error!()'s, it is possible to configure log-based alarms so that these failures can be caught quickly, but that is outside the scope of this README.

Tracing

This crate emits trace data tracing and can be integrated into standard tracing implementations.

Authorization Logging

Authorization logs are designed to power detection and response capabilities. Sample capabilities can be found under the Mitre D3fend matrix, for example User Behavioral Analysis.

The authorizer's provided emit authorization events as tracing events. Authorization events are included in tracing spans. Authorization events are default formatted using Open Cyber Security Format. Authorization events can optionally be filtered, formatted and routed directly to an authorization log. See example:

// Dependencies must be included in the `Cargo.toml` file of the application
// tracing, tracing-appender, tracing-subscriber

let authorization_roller = tracing_appender::rolling::minutely("logs", "authorization.log");
let (authorization_non_blocking, _guard) = tracing_appender::non_blocking(authorization_roller);
let authorization_log_layer = tracing_subscriber::fmt::layer()
    .event_format(AuthorizerFormatter(AuthorizerTarget::Simple))
    .with_writer(authorization_non_blocking);
tracing_subscriber::registry()
    .with(authorization_log_layer)
    .try_init()
    .expect("Logging Failed to Start, Exiting.");

To filter authorization event logs, provide a log config to the authorizer with a FieldSet which includes the fields that are to be logged. By default if not explicitly configured, no fields will be logged.

Sample usage of logging everything within the authorization request. Note: Logging everything is insecure; please see Secure Logging Configuration.

let log_config =
    log::ConfigBuilder::default()
        .field_set(log::FieldSetBuilder::default()
            .principal(true)
            .action(true)
            .resource(true)
            .context(true)
            .entities(log::FieldLevel::All)
            .build()
            .unwrap())
    .build()
    .unwrap();

let authorizer: Authorizer<PolicySetProvider, EntityProvider> = Authorizer::new(
    AuthorizerConfigBuilder::default()
        .entity_provider(...)
        .policy_set_provider(...)
        .log_config(log_config)
        .build()
        .unwrap(),
);

Note: Authorizer log_config FieldSet::entities refers to the Cedar entities. There is also an OCSF field called entity which refers to the principal entity that is sending the request.

This means that when Authorizer log_config FieldSet::entities is set to FieldLevel::None, the OCSF entity will still be logged. This is not a bug and is expected behaviour.

For more examples of how to set up the authorization logging, see our usage examples

Secure Logging Configuration:

Using a log::FieldSet configuration that sets any cedar-related field (principal, action, resource, context, and entities) to true will result in that field being logged. These fields could contain sensitive information and should be exposed with caution. Additionally, the cedar language has no current limit on field sizes within a Request. A large request with verbose logging can result in more disk i/o to occur. This disk i/o could negatively impact the performance of the application.

For a safe config, create a default log::FieldSet with log::FieldSetBuilder::default().build().unwrap(). This option will redact user input like so:

"entity":{"data":{"Parents":[]},"name":"Sensitive<REDACTED>","type":"Sensitive<REDACTED>"}

Note:

Cedar does not at this time support extracting the Context from the Request struct since it is private, therefore it is extracted using request.to_string(). This is not ideal as this logs the entire request (Principal, Action, Resource, Context) instead of just the context.

In particular this brings two quirks:

  • If the FieldSet has principal and context set to true, then the resulting log will include the principal twice.
  • If the FieldSet has principal set to false and context set to true, then the resulting log will include principal anyway since it is included in the request.to_string() call required to extract the context.

The above are not specific to principal and also occur with action and resource. A cedar github issue has been created to add a getter for the context on the cedar Request struct that will fix this: cedar-policy/cedar#363

Example application

This project is based on a fictitious application that allows users to manage sweet boxes. There are two entities:

  1. Box (Resource)
  2. User (Principal)

There are two user entities:

  1. Eric
  2. Mike

There are ten resources, Box. Each box has an entity identifier (id) that ranges between the numbers 1-10. Each Box has one attribute owner. The owners of each Box are defined in the tests/data/sweets.entities.json file, this file represents the complete database of information for this application.

The actions that Users can perform on each Box are defined in the schema file: tests/data/sweets.schema.cedar.json. To summarize, each User can perform one of the following actions read, update or delete on a Box resource.

A policy is a statement that declares which Users are explicitly permitted, or explicitly forbidden to perform an action on a resource, Box. Here is a sample policy:

@id("owner-policy")
permit(principal, action, resource)
when { principal == resource.owner };

Refer to tests/data/sweets.cedar for the full details of these policies. This file represents all policies for this application.

Given the schema, entities and policy_set the application can use the Authorizer as provided in the usage above. Here is a sample request and expected outcome:

 assert_eq!(
     authorizer
         .is_authorized(&Request::new(
             Some(format!("User::\"Mike\"").parse().unwrap()),
             Some(format!("Action::\"read\"").parse().unwrap()),
             Some(format!("Box::\"3\"").parse().unwrap()),
             Context::empty(),
         ), &Entities::empty())
         .await
         .unwrap()
         .decision(),
     Decision::Deny
 );
 assert_eq!(
     authorizer
         .is_authorized(&Request::new(
             Some(format!("User::\"Mike\"").parse().unwrap()),
             Some(format!("Action::\"read\"").parse().unwrap()),
             Some(format!("Box::\"2\"").parse().unwrap()),
             Context::empty(),
         ), &Entities::empty())
         .await
         .unwrap()
         .decision(),
     Decision::Allow
 );

Feel free to refer to this sample application within the integration test located here: tests/lib.rs.

General Security Notes

The following is a high level description of some security concerns to keep in mind when using the cedar-local-agent to enable local evaluation of Cedar policies stored on a local file system.

Trusted Computing Environment

The cedar-local-agent is a mere library that customers can wrap in say an HTTP server and deploy onto a fleet of hosts. It is, therefore, left to users to take any and all necessary precautions to ensure those security concerns beyond what the cedar-local-agent is capable of enforcing are met. This includes:

  1. Filesystem permissions for on-disk Policy Stores should be limited to least-privilege, see Limiting Access to Local Data Files.
  2. Filesystem permissions for on-disk locations of OCSF logs follow least-privilege permissions, see OCSF Log directory permissions.
  3. The cedar-local-agent is configured securely, see Quick Start and Updating file::PolicySetProvider or file::EntityProvider data for configuration best practices.
  4. Validating the size of files read and requests made to is_authorized to prevent potential denial of service.

Limiting Access to Local Data Files

The local authorizer provided in this crate only needs read access to locally stored policy set, entity store and schema files.

Write access to local data files (policies, entities and schema) should be restricted only to users that really need to make changes to these files, for example, to add new entities and remove old policies.

In the case where there are no restrictions to access local data files, a malicious Operating System (OS) user can add or remove policies, modify entities attributes, make slight changes that are hard to identify, or even change the policies to deny all actions. To illustrate this possibility, consider a cedar file with the following cedar policies from the [Example Application](## Example application):

@id("mike-edit-box-1")
permit (
    principal == User::"Mike",
    action == Action::"update",
    resource == Box::"1"
);

@id("eric-view-box-9")
permit (
    principal == User::"Eric",
    action == Action::"read",
    resource == Box::"9"
);

In this example, principal "Mike" is allowed to perform "update" on resource box "1" while principal "Eric" is allowed to perform "read" on resource box "9". Now, consider a malicious OS user adding the statement below to the same policies file.

@id("ill-intentioned-policy")
forbid(principal, action, resource);

In the next policies file refresh cycle, the file::PolicySetProvider will refresh policies file content to memory, and the local authorizer will deny any action from any principal.

How to avoid this problem from happening?

In order to prevent this kind of security issue, you must restrict read access to the data files, and more important, restrict write access to these files. Only users or groups that really need to write changes to policies, or entities should be allowed to do so (for example, another agent that fetches policies from an internal application).

For one example on how to avoid this problem, say you have the following folder structure for a local-agent built with cedar-local-agent crate.

authz-agent/
  |- authz_daemon (executable)

authz-local-data/
  |- policies.cedar
  |- entities.json
  |- schema.json

Now suppose you have an OS user to execute the "authz_daemon" called "authz-daemon" from user group "authz-ops". And you have a user called "authz-ops-admin" from the same user group "authz-ops" that will be able to update data files.

Then, make "authz-ops-admin" the owner of authz-local-data folder with:

$ chown -R authz-ops-admin:authz-ops authz-local-data

And make "authz-daemon" user the owner of authz-agent folder with:

$ chown -R authz-daemon:authz-ops authz-agent

Finally, make authz-local-data readable by everyone and writable by the owner only:

$ chmod u=rwx,go=r authz-local-data

OCSF Log directory permissions

The local authorizer provided in this crate will require read and write access to the directory where it will write OCFS logs to.

Suppose we have the following directory structure:

authz-agent/
  |- authz_daemon (executable)

ocsf-log-dir/
  |- authorization.log.2023-11-15-21-02
  ...

Now suppose you have an OS user to execute the authz_daemon called authz-daemon which should be in a group called "log-reader".

And make authz-daemon user the owner of ocsf-log-dir folder with:

$ chown -R authz-daemon:log-reader ocsf-log-dir

We will now make ocsf-log-dir readable and writable by the owner but not writable to anyone else. We allow anyone in the log-reader group to read the contents of the folder but not write to it.

$ chmod u=wrx,g=rx,o= ocsf-log-dir

NOTE: We need to allow execute permissions in order to access files in the directory.

Any agent that needs to access the logs, such as the AWS Cloudwatch Agent should run as a user in the log-reader group so that they will have the proper access (see documentation for how to configure the Cloudwatch Agent to run as a certain user).

Getting Help

License

This project is licensed under the Apache-2.0 License.

About

No description, website, or topics provided.

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Rust 98.9%
  • Shell 1.1%