This package contains static libraries: libpulsar.a, libpulsarwithdeps.a and C/C++ headers.
Install the package using the following command:
rpm-ivh apache-pulsar-client*.rpm
Now, you can see Pulsar C++ client libraries installed under the /usr/lib directory.
note
If you get an error like "libpulsar.so: cannot open shared object file: No such file or directory" when starting a Pulsar client, you need to run ldconfig first.
To connect Pulsar using client libraries, you need to specify a Pulsar protocol URL.
Pulsar protocol URLs are assigned to specific clusters, you can use the Pulsar URI scheme. The default port is 6650. The following is an example for localhost.
pulsar://localhost:6650
In a Pulsar cluster in production, the URL looks as follows.
pulsar://pulsar.us-west.example.com:6650
If you use TLS authentication, you need to add ssl, and the default port is 6651. The following is an example.
This example sends 100 messages using the blocking style. While simple, it does not produce high throughput as it waits for each ack to come back before sending the next message.
#include<pulsar/Client.h> #include<thread> usingnamespace pulsar; intmain(){ Client client("pulsar://localhost:6650"); Result result = client.createProducer("persistent://public/default/my-topic", producer); if(result != ResultOk){ std::cout <<"Error creating producer: "<< result << std::endl; return-1; } // Send 100 messages synchronously int ctr =0; while(ctr <100){ std::string content ="msg"+ std::to_string(ctr); Message msg =MessageBuilder().setContent(content).setProperty("x","1").build(); Result result = producer.send(msg); if(result != ResultOk){ std::cout <<"The message "<< content <<" could not be sent, received code: "<< result << std::endl; }else{ std::cout <<"The message "<< content <<" sent successfully"<< std::endl; } std::this_thread::sleep_for(std::chrono::milliseconds(100)); ctr++; } std::cout <<"Finished producing synchronously!"<< std::endl; client.close(); return0; }
This example sends 100 messages using the non-blocking style calling sendAsync instead of send. This allows the producer to have multiple messages inflight at a time which increases throughput.
The producer configuration blockIfQueueFull is useful here to avoid ResultProducerQueueIsFull errors when the internal queue for outgoing send requests becomes full. Once the internal queue is full, sendAsync becomes blocking which can make your code simpler.
Without this configuration, the result code ResultProducerQueueIsFull is passed to the callback. You must decide how to deal with that (retry, discard etc).
When scaling out a Pulsar topic, you may configure a topic to have hundreds of partitions. Likewise, you may have also scaled out your producers so there are hundreds or even thousands of producers. This can put some strain on the Pulsar brokers as when you create a producer on a partitioned topic, internally it creates one internal producer per partition which involves communications to the brokers for each one. So for a topic with 1000 partitions and 1000 producers, it ends up creating 1,000,000 internal producers across the producer applications, each of which has to communicate with a broker to find out which broker it should connect to and then perform the connection handshake.
You can reduce the load caused by this combination of a large number of partitions and many producers by doing the following:
use SinglePartition partition routing mode (this ensures that all messages are only sent to a single, randomly selected partition)
use non-keyed messages (when messages are keyed, routing is based on the hash of the key and so messages will end up being sent to multiple partitions)
use lazy producers (this ensures that an internal producer is only created on demand when a message needs to be routed to a partition)
With our example above, that reduces the number of internal producers spread out over the 1000 producer apps from 1,000,000 to just 1000.
Note that there can be extra latency for the first message sent. If you set a low send timeout, this timeout could be reached if the initial connection handshake is slow to complete.
Message chunking enables Pulsar to process large payload messages by splitting the message into chunks at the producer side and aggregating chunked messages at the consumer side.
The message chunking feature is OFF by default. The following is an example about how to enable message chunking when creating a producer.
You can avoid running a loop by blocking calls with an event-based style by using a message listener which is invoked for each message that is received.
This example starts a subscription at the earliest offset and consumes 100 messages.
#include<pulsar/Client.h> #include<atomic> #include<thread> usingnamespace pulsar; std::atomic<uint32_t> messagesReceived; voidhandleAckComplete(Result res){ std::cout <<"Ack res: "<< res << std::endl; } voidlistener(Consumer consumer,const Message& msg){ std::cout <<"Got message "<< msg <<" with content '"<< msg.getDataAsString()<<"'"<< std::endl; messagesReceived++; consumer.acknowledgeAsync(msg.getMessageId(), handleAckComplete); } intmain(){ Client client("pulsar://localhost:6650"); Consumer consumer; ConsumerConfiguration config; config.setMessageListener(listener); config.setSubscriptionInitialPosition(InitialPositionEarliest); Result result = client.subscribe("persistent://public/default/my-topic","consumer-1", config, consumer); if(result != ResultOk){ std::cout <<"Failed to subscribe: "<< result << std::endl; return-1; } // wait for 100 messages to be consumed while(messagesReceived <100){ std::this_thread::sleep_for(std::chrono::milliseconds(100)); } std::cout <<"Finished consuming asynchronously!"<< std::endl; client.close(); return0; }
You can limit the maximum number of chunked messages a consumer maintains concurrently by configuring the setMaxPendingChunkedMessage and setAutoAckOldestChunkedMessageOnQueueFull parameters. When the threshold is reached, the consumer drops pending messages by silently acknowledging them or asking the broker to redeliver them later.
The following is an example of how to configure message chunking.
If you use TLS authentication when connecting to Pulsar, you need to add ssl in the connection URLs, and the default port is 6651. The following is an example.