Confluent Certified Cloud Operator (CCAC)
- 60 multiple choice questions in 90 minutes
- The Certification expires after two years
Product Knowledge
Understanding of:
- Confluent Cloud core concepts, including clusters, client applications, connectors, Stream Governance, and stream processing
- Basic Kafka concepts, Kafka Connect and Operations
- Confluent Cloud security features/offerings
- Cloud architecture, networking and computing
- Monitoring and alerting
| Certification Outline | Break Down by % |
|---|---|
| Confluent Cloud Core Concepts | 17% |
| Kafka Operations | 17% |
| Confluent Cloud Static Operations | 14% |
| Confluent Cloud Dynamic Operations | 16% |
| Confluent Cloud Streaming Pipelines | 11% |
| Confluent Cloud Data Governance | 13% |
| Confluent Cloud Resilience | 11% |
| 100% |
- Cluster types - Basic, Standard, Enterprise and dedicated Confluent Cloud cluster setup.
- Understand accounts, API Keys and authorization for Confluent Cloud.
- Configure Managed connectors
- OIDC provider to authenticate services
- Understanding Audit Logs and Metrics
- Audit logs via cluster linking
- Cluster and schema linking
- Confluent Terraform provider for Infrastructure as Code
- Private Networking
- Schema Governance
- Authentication and Authorization
- Managed Connectors
- Audit logs and metrics
Concepts
- Confluent Cloud Core Concepts
- Differentiate between billable and non-billable cost factors within Confluent Cloud
- Differentiate between cluster types including features and limitations
- Describe the relationship between different Confluent Cloud resources
- List the components and resources that require public networking
- Compare responsibilities of the Confluent Cloud role with the self-managed OSK role
- Kafka Operations
- Match Kafka keywords to their functions in the Kafka ecosystem
- Classify which cluster configurations are and are not exposed in Confluent Cloud
- Classify which topic configurations are and are not exposed in Confluent Cloud
- Describe the required elements needed to connect to a cluster
- Define the key configuration properties associated with topic creation
- Understand how to change topic configurations throughout the topic lifecycle
- Confluent Cloud Static Operations
- Identify available networking options to provision a cluster
- List the different ways to provide access to all UI features with private networking
- Describe the options when provisioning a cluster
- Define the purpose of an environment
- Design integration with the Metrics API
- Design the method of capturing audit logs
- Define the operations that are captured in the audit log
- List the tooling available to provision Confluent Cloud resources
- Confluent Cloud Dynamic Operations
- Differentiate Cloud API keys and Cluster API keys
- Compare and contrast users and service accounts
- Evaluate Terraform/CLI/REST API vs UI for deploying resources
- List the best practices for securely onboarding applications
- Compare RBAC vs ACLs
- Differentiate Confluent Cloud authentication options for users and service accounts
- Assess the importance of least privileges for users and applications
- Confluent Cloud Streaming Pipelines
- Define the relationship between Connectors and external systems
- Describe how Connector tasks relate to performance
- Determine when fully-managed connectors can be used and when it's necessary to use self-managed connectors
- Define typical requirements to provision a Connector
- List the differences between provisioning a ksqlDB cluster and Flink compute pool
- Compare application capabilities when using ksqlDB vs Flink
- Confluent Cloud Data Governance
- Define the purpose of schemas in Confluent Cloud and how to use them
- List the methods of managing the lifecycle of schemas
- Match the Stream Governance feature to its purpose
- Compare Stream Governance packages
- Confluent Cloud Resilience
- Evaluate scenarios when cluster linking and/or schema linking should be used
- Differentiate High Availability use cases from Disaster Recovery use-cases
- Classify use-cases for DR Readiness
- List the benefits of using schema linking in synchronizing data
- List the benefits of using cluster linking in synchronizing data
- List HA configurations for non-Kafka and SR components
Questions
- A production Kafka cluster in Confluent Cloud is experiencing consumer lag issues. Which of the following is the most effective first step to identify the root cause of this performance problem?
- Immediately increase the number of partitions for all topics to improve parallelism
- Review consumer group metrics and partition assignment distribution in the Confluent Cloud monitoring dashboard
- Restart all consumer applications to reset their connection to the Kafka brokers
- Increase the replication factor for all topics to improve fault tolerance
- Your security team requires all production applications be deployed with AWS Private Link networking. Which two Kafka cluster types could be used to meet this requirement?
- Basic
- Standard
- Enterprise
- Dedicated
- Advanced
- You are creating a new topic for an application which will have several concurrent consumers in a single consumer group. What configuration setting will you need to carefully tune in order to meet this requirement?
- num.partitions
- retention.ms
- cleanup.policy
- retention.bytes
- flush.ms
- Your organization’s usage of Confluent Cloud has grown and onboarding new user accounts has been error prone and adds to the administrative burden on the operations team. What changes can you make to streamline onboarding new users and simplify identity management for Confluent Cloud?
- Enable Single sign-on (SSO) on Confluent Cloud for your organization’s identity provider
- Have each team create a Confluent Cloud organization and allow them to manage their own organization users
- Create service accounts for each team to share
- Create one user/password pair for each team and let the team share a single user to access Confluent Cloud.
- What will happen if a Kafka producer sends data with acks=0?
- The message is only stored in memory
- No acknowledgment is expected; potential data loss
- The broker guarantees delivery
- Messages are written to all replicas
- What happens if you attempt to create a network connection in a region not supported by your selected cloud provider?
- The CLI returns an error
- It silently fails
- It defaults to multi-region
- The connection is automatically rerouted
- You observe that consumers are experiencing increased lag. What is the most likely impact?
- Data loss
- Consumers read duplicate messages
- Consumers take longer to process new messages
- Producers are throttled
- Which cloud-specific resource must be configured when enabling Private Link for Confluent Cloud on AWS?
- Cloud NAT Gateway
- VPC Endpoint Service
- Internet Gateway
- Route 53 Resolver
- In Confluent Cloud, which two components are included in the Stream Governance “Advanced” tier? (Choose two)
- Stream Catalog
- Audit Logs
- Stream Lineage
- Connectors
- You are planning to organize multiple projects for different teams. What two features should you use to isolate and manage their access? (Choose two)
- Role Bindings
- Multiple Clusters
- Environments
- Regions
- Which two statements accurately describe how producers interact with partitions in Kafka? (Choose two)
- Producers send all messages to partition 0 by default
- A key ensures messages go to the same partition
- Producers manually assign offsets
- Partitioning helps achieve parallelism in message delivery
- When creating a Kafka topic in Confluent Cloud, what happens if no replication factor is specified?
- It uses the replication factor of the cluster configuration
- It depends on the environment
- It defaults to 1
- You must manually specify it; otherwise topic creation fails
- What is the primary purpose of configuring network access on a Kafka cluster in Confluent Cloud?
- To optimize message compression settings
- To enable multiple schema versions
- To restrict or allow client access via public or private endpoints
- To limit partition replication
- Which ksqlDB construct should be used for building materialized views of the latest value per key?
- STREAM
- SCHEMA
- JOIN
- TABLE
- You are tasked with designing a multi-region deployment for a streaming application on Confluent Cloud. The application must maintain low latency for users across different geographical locations while ensuring high availability and data consistency. What architectural considerations should you prioritize when setting up your Kafka clusters in Confluent Cloud to meet these requirements?
- Deploy a single Kafka cluster in the closest region to the majority of users and use partitioning to manage data distribution.
- Set up multiple Kafka clusters in each region and use Confluent Replicator to ensure data consistency across them.
- Implement a multi-region Kafka cluster configuration in Confluent Cloud, leveraging partition leaders in the respective regions to minimize latency while using global data replication for consistency.
- Utilize a single Kafka cluster with a geo-replication strategy but limit the number of partitions to reduce overhead.
- You are tasked with provisioning a new Kafka cluster in Confluent Cloud to support a high-throughput application with strict latency requirements. The application is expected to handle over a million messages per second and requires low-latency processing. Which configuration should you choose to ensure optimal performance while maintaining cost-effectiveness?
- Provision a cluster with the lowest available instance type to minimize costs.
- Select a cluster with multiple high-capacity brokers but limit the number of partitions to minimize resource usage.
- Provision a cluster with high-capacity brokers and an appropriate number of partitions to maximize parallelism.
- Choose a multi-region cluster to ensure redundancy, sacrificing some performance for availability.
- You are managing a Kafka cluster in Confluent Cloud with a replication factor of 3 for your topics. Recently, one of the brokers went down, and you noticed that some partitions have not reached the ISR (In-Sync Replica) state. After the broker is restored, you observe that some messages are missing from the topic. How should you handle this situation to ensure data consistency and availability for your consumers?
- Increase the replication factor to 5 for the affected topic and recreate the partitions.
- Enable the log retention policy to delete older messages and allow the topic to catch up.
- Manually trigger a leader election for the affected partitions to ensure that the remaining in-sync replicas can take over.
- This option is CORRECT because triggering a leader election allows other in-sync replicas to become leaders, which can help restore availability and ensure consumers can access the latest data.
- Reconfigure the replication factor to 1 temporarily to improve performance while the broker is down.
- You are managing a Confluent Cloud environment for a financial services company that deals with sensitive customer data. Due to regulatory compliance requirements, you need to restrict access to specific Kafka topics based on user roles. The team debates whether to use Access Control Lists (ACLs) or Role-Based Access Control (RBAC) for this purpose. Given the context, which approach is more suitable and why?
- Use ACLs to grant permissions to individual users for each topic.
- Implement RBAC to define roles that can access multiple topics based on job functions.
- Utilize RBAC to create roles that are tied to regulatory compliance requirements, ensuring users can only access topics relevant to their roles.
- Combine ACLs and RBAC to handle both individual permissions and role-based access.
- You are tasked with integrating a legacy SQL database with a real-time analytics platform using Kafka Connect in Confluent Cloud. The database has a large volume of historical data and frequent updates. After configuring the JDBC source connector, you notice that the initial data load is taking an unusually long time, and you are concerned about the impact on the performance of your Kafka cluster and downstream consumers. What is the best approach to optimize the initial data load while ensuring minimal disruption to the existing data pipeline
- Increase the 'batch.size' configuration in the connector settings to load more records at once.
- Set the poll interval.ms' to a lower value to increase the frequency of data fetching from the database.
- Use the 'incrementing.column.name' and 'incrementing offset properties' to perform the initial load in smaller chunks based on the last update timestamps.
- Run the initial load during off-peak hours to minimize the impact on the performance of the Kafka cluster and downstream consumers.