2013-01-20 11:18:40 +01:00
|
|
|
/*
|
|
|
|
* Copyright 2013 The Netty Project
|
|
|
|
*
|
|
|
|
* The Netty Project licenses this file to you under the Apache License,
|
|
|
|
* version 2.0 (the "License"); you may not use this file except in compliance
|
|
|
|
* with the License. You may obtain a copy of the License at:
|
|
|
|
*
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
*
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
|
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
|
|
* License for the specific language governing permissions and limitations
|
|
|
|
* under the License.
|
|
|
|
*/
|
|
|
|
package io.netty.channel.rxtx;
|
|
|
|
|
|
|
|
import io.netty.buffer.ByteBufAllocator;
|
|
|
|
import io.netty.channel.ChannelOption;
|
|
|
|
import io.netty.channel.DefaultChannelConfig;
|
Revamp the core API to reduce memory footprint and consumption
The API changes made so far turned out to increase the memory footprint
and consumption while our intention was actually decreasing them.
Memory consumption issue:
When there are many connections which does not exchange data frequently,
the old Netty 4 API spent a lot more memory than 3 because it always
allocates per-handler buffer for each connection unless otherwise
explicitly stated by a user. In a usual real world load, a client
doesn't always send requests without pausing, so the idea of having a
buffer whose life cycle if bound to the life cycle of a connection
didn't work as expected.
Memory footprint issue:
The old Netty 4 API decreased overall memory footprint by a great deal
in many cases. It was mainly because the old Netty 4 API did not
allocate a new buffer and event object for each read. Instead, it
created a new buffer for each handler in a pipeline. This works pretty
well as long as the number of handlers in a pipeline is only a few.
However, for a highly modular application with many handlers which
handles connections which lasts for relatively short period, it actually
makes the memory footprint issue much worse.
Changes:
All in all, this is about retaining all the good changes we made in 4 so
far such as better thread model and going back to the way how we dealt
with message events in 3.
To fix the memory consumption/footprint issue mentioned above, we made a
hard decision to break the backward compatibility again with the
following changes:
- Remove MessageBuf
- Merge Buf into ByteBuf
- Merge ChannelInboundByte/MessageHandler and ChannelStateHandler into ChannelInboundHandler
- Similar changes were made to the adapter classes
- Merge ChannelOutboundByte/MessageHandler and ChannelOperationHandler into ChannelOutboundHandler
- Similar changes were made to the adapter classes
- Introduce MessageList which is similar to `MessageEvent` in Netty 3
- Replace inboundBufferUpdated(ctx) with messageReceived(ctx, MessageList)
- Replace flush(ctx, promise) with write(ctx, MessageList, promise)
- Remove ByteToByteEncoder/Decoder/Codec
- Replaced by MessageToByteEncoder<ByteBuf>, ByteToMessageDecoder<ByteBuf>, and ByteMessageCodec<ByteBuf>
- Merge EmbeddedByteChannel and EmbeddedMessageChannel into EmbeddedChannel
- Add SimpleChannelInboundHandler which is sometimes more useful than
ChannelInboundHandlerAdapter
- Bring back Channel.isWritable() from Netty 3
- Add ChannelInboundHandler.channelWritabilityChanges() event
- Add RecvByteBufAllocator configuration property
- Similar to ReceiveBufferSizePredictor in Netty 3
- Some existing configuration properties such as
DatagramChannelConfig.receivePacketSize is gone now.
- Remove suspend/resumeIntermediaryDeallocation() in ByteBuf
This change would have been impossible without @normanmaurer's help. He
fixed, ported, and improved many parts of the changes.
2013-05-28 13:40:19 +02:00
|
|
|
import io.netty.channel.RecvByteBufAllocator;
|
2013-01-20 11:18:40 +01:00
|
|
|
|
|
|
|
import java.util.Map;
|
|
|
|
|
|
|
|
import static io.netty.channel.rxtx.RxtxChannelOption.*;
|
|
|
|
|
|
|
|
/**
|
|
|
|
* Default configuration class for RXTX device connections.
|
|
|
|
*/
|
|
|
|
final class DefaultRxtxChannelConfig extends DefaultChannelConfig implements RxtxChannelConfig {
|
|
|
|
|
|
|
|
private volatile int baudrate = 115200;
|
|
|
|
private volatile boolean dtr;
|
|
|
|
private volatile boolean rts;
|
|
|
|
private volatile Stopbits stopbits = Stopbits.STOPBITS_1;
|
|
|
|
private volatile Databits databits = Databits.DATABITS_8;
|
|
|
|
private volatile Paritybit paritybit = Paritybit.NONE;
|
2013-01-21 15:12:23 +01:00
|
|
|
private volatile int waitTime;
|
2013-06-10 18:56:33 +02:00
|
|
|
private volatile int readTimeout = 1000;
|
2013-01-20 11:18:40 +01:00
|
|
|
|
|
|
|
public DefaultRxtxChannelConfig(RxtxChannel channel) {
|
|
|
|
super(channel);
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public Map<ChannelOption<?>, Object> getOptions() {
|
2013-01-21 15:12:23 +01:00
|
|
|
return getOptions(super.getOptions(), BAUD_RATE, DTR, RTS, STOP_BITS, DATA_BITS, PARITY_BIT, WAIT_TIME);
|
2013-01-20 11:18:40 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
@SuppressWarnings("unchecked")
|
|
|
|
@Override
|
|
|
|
public <T> T getOption(ChannelOption<T> option) {
|
|
|
|
if (option == BAUD_RATE) {
|
|
|
|
return (T) Integer.valueOf(getBaudrate());
|
|
|
|
}
|
|
|
|
if (option == DTR) {
|
|
|
|
return (T) Boolean.valueOf(isDtr());
|
|
|
|
}
|
|
|
|
if (option == RTS) {
|
|
|
|
return (T) Boolean.valueOf(isRts());
|
|
|
|
}
|
|
|
|
if (option == STOP_BITS) {
|
|
|
|
return (T) getStopbits();
|
|
|
|
}
|
|
|
|
if (option == DATA_BITS) {
|
|
|
|
return (T) getDatabits();
|
|
|
|
}
|
|
|
|
if (option == PARITY_BIT) {
|
|
|
|
return (T) getParitybit();
|
|
|
|
}
|
2013-01-21 15:12:23 +01:00
|
|
|
if (option == WAIT_TIME) {
|
|
|
|
return (T) Integer.valueOf(getWaitTimeMillis());
|
|
|
|
}
|
2013-06-10 18:56:33 +02:00
|
|
|
if (option == READ_TIMEOUT) {
|
|
|
|
return (T) Integer.valueOf(getReadTimeout());
|
|
|
|
}
|
2013-01-20 11:18:40 +01:00
|
|
|
return super.getOption(option);
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public <T> boolean setOption(ChannelOption<T> option, T value) {
|
|
|
|
validate(option, value);
|
|
|
|
|
|
|
|
if (option == BAUD_RATE) {
|
|
|
|
setBaudrate((Integer) value);
|
|
|
|
} else if (option == DTR) {
|
|
|
|
setDtr((Boolean) value);
|
|
|
|
} else if (option == RTS) {
|
|
|
|
setRts((Boolean) value);
|
|
|
|
} else if (option == STOP_BITS) {
|
|
|
|
setStopbits((Stopbits) value);
|
|
|
|
} else if (option == DATA_BITS) {
|
|
|
|
setDatabits((Databits) value);
|
|
|
|
} else if (option == PARITY_BIT) {
|
|
|
|
setParitybit((Paritybit) value);
|
2013-01-21 15:12:23 +01:00
|
|
|
} else if (option == WAIT_TIME) {
|
|
|
|
setWaitTimeMillis((Integer) value);
|
2013-06-10 18:56:33 +02:00
|
|
|
} else if (option == READ_TIMEOUT) {
|
|
|
|
setReadTimeout((Integer) value);
|
2013-01-20 11:18:40 +01:00
|
|
|
} else {
|
|
|
|
return super.setOption(option, value);
|
|
|
|
}
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setBaudrate(final int baudrate) {
|
|
|
|
this.baudrate = baudrate;
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setStopbits(final Stopbits stopbits) {
|
|
|
|
this.stopbits = stopbits;
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setDatabits(final Databits databits) {
|
|
|
|
this.databits = databits;
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setParitybit(final Paritybit paritybit) {
|
|
|
|
this.paritybit = paritybit;
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public int getBaudrate() {
|
|
|
|
return baudrate;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public Stopbits getStopbits() {
|
|
|
|
return stopbits;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public Databits getDatabits() {
|
|
|
|
return databits;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public Paritybit getParitybit() {
|
|
|
|
return paritybit;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public boolean isDtr() {
|
|
|
|
return dtr;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setDtr(final boolean dtr) {
|
|
|
|
this.dtr = dtr;
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public boolean isRts() {
|
|
|
|
return rts;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setRts(final boolean rts) {
|
|
|
|
this.rts = rts;
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
2013-01-21 15:12:23 +01:00
|
|
|
@Override
|
|
|
|
public int getWaitTimeMillis() {
|
|
|
|
return waitTime;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setWaitTimeMillis(final int waitTimeMillis) {
|
|
|
|
if (waitTimeMillis < 0) {
|
|
|
|
throw new IllegalArgumentException("Wait time must be >= 0");
|
|
|
|
}
|
|
|
|
waitTime = waitTimeMillis;
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
2013-06-10 18:56:33 +02:00
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setReadTimeout(int readTimeout) {
|
|
|
|
if (readTimeout < 0) {
|
|
|
|
throw new IllegalArgumentException("readTime must be >= 0");
|
|
|
|
}
|
|
|
|
this.readTimeout = readTimeout;
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public int getReadTimeout() {
|
|
|
|
return readTimeout;
|
|
|
|
}
|
|
|
|
|
2013-01-20 11:18:40 +01:00
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setConnectTimeoutMillis(int connectTimeoutMillis) {
|
2013-06-25 10:49:28 +02:00
|
|
|
super.setConnectTimeoutMillis(connectTimeoutMillis);
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setMaxMessagesPerRead(int maxMessagesPerRead) {
|
|
|
|
super.setMaxMessagesPerRead(maxMessagesPerRead);
|
|
|
|
return this;
|
2013-01-20 11:18:40 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setWriteSpinCount(int writeSpinCount) {
|
2013-06-25 10:49:28 +02:00
|
|
|
super.setWriteSpinCount(writeSpinCount);
|
|
|
|
return this;
|
2013-01-20 11:18:40 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setAllocator(ByteBufAllocator allocator) {
|
2013-06-25 10:49:28 +02:00
|
|
|
super.setAllocator(allocator);
|
|
|
|
return this;
|
2013-01-20 11:18:40 +01:00
|
|
|
}
|
|
|
|
|
Revamp the core API to reduce memory footprint and consumption
The API changes made so far turned out to increase the memory footprint
and consumption while our intention was actually decreasing them.
Memory consumption issue:
When there are many connections which does not exchange data frequently,
the old Netty 4 API spent a lot more memory than 3 because it always
allocates per-handler buffer for each connection unless otherwise
explicitly stated by a user. In a usual real world load, a client
doesn't always send requests without pausing, so the idea of having a
buffer whose life cycle if bound to the life cycle of a connection
didn't work as expected.
Memory footprint issue:
The old Netty 4 API decreased overall memory footprint by a great deal
in many cases. It was mainly because the old Netty 4 API did not
allocate a new buffer and event object for each read. Instead, it
created a new buffer for each handler in a pipeline. This works pretty
well as long as the number of handlers in a pipeline is only a few.
However, for a highly modular application with many handlers which
handles connections which lasts for relatively short period, it actually
makes the memory footprint issue much worse.
Changes:
All in all, this is about retaining all the good changes we made in 4 so
far such as better thread model and going back to the way how we dealt
with message events in 3.
To fix the memory consumption/footprint issue mentioned above, we made a
hard decision to break the backward compatibility again with the
following changes:
- Remove MessageBuf
- Merge Buf into ByteBuf
- Merge ChannelInboundByte/MessageHandler and ChannelStateHandler into ChannelInboundHandler
- Similar changes were made to the adapter classes
- Merge ChannelOutboundByte/MessageHandler and ChannelOperationHandler into ChannelOutboundHandler
- Similar changes were made to the adapter classes
- Introduce MessageList which is similar to `MessageEvent` in Netty 3
- Replace inboundBufferUpdated(ctx) with messageReceived(ctx, MessageList)
- Replace flush(ctx, promise) with write(ctx, MessageList, promise)
- Remove ByteToByteEncoder/Decoder/Codec
- Replaced by MessageToByteEncoder<ByteBuf>, ByteToMessageDecoder<ByteBuf>, and ByteMessageCodec<ByteBuf>
- Merge EmbeddedByteChannel and EmbeddedMessageChannel into EmbeddedChannel
- Add SimpleChannelInboundHandler which is sometimes more useful than
ChannelInboundHandlerAdapter
- Bring back Channel.isWritable() from Netty 3
- Add ChannelInboundHandler.channelWritabilityChanges() event
- Add RecvByteBufAllocator configuration property
- Similar to ReceiveBufferSizePredictor in Netty 3
- Some existing configuration properties such as
DatagramChannelConfig.receivePacketSize is gone now.
- Remove suspend/resumeIntermediaryDeallocation() in ByteBuf
This change would have been impossible without @normanmaurer's help. He
fixed, ported, and improved many parts of the changes.
2013-05-28 13:40:19 +02:00
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setRecvByteBufAllocator(RecvByteBufAllocator allocator) {
|
|
|
|
super.setRecvByteBufAllocator(allocator);
|
|
|
|
return this;
|
|
|
|
}
|
|
|
|
|
2013-01-20 11:18:40 +01:00
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setAutoRead(boolean autoRead) {
|
2013-06-25 10:49:28 +02:00
|
|
|
super.setAutoRead(autoRead);
|
|
|
|
return this;
|
2013-01-20 11:18:40 +01:00
|
|
|
}
|
2013-03-08 08:20:46 +01:00
|
|
|
|
|
|
|
@Override
|
Revamp the core API to reduce memory footprint and consumption
The API changes made so far turned out to increase the memory footprint
and consumption while our intention was actually decreasing them.
Memory consumption issue:
When there are many connections which does not exchange data frequently,
the old Netty 4 API spent a lot more memory than 3 because it always
allocates per-handler buffer for each connection unless otherwise
explicitly stated by a user. In a usual real world load, a client
doesn't always send requests without pausing, so the idea of having a
buffer whose life cycle if bound to the life cycle of a connection
didn't work as expected.
Memory footprint issue:
The old Netty 4 API decreased overall memory footprint by a great deal
in many cases. It was mainly because the old Netty 4 API did not
allocate a new buffer and event object for each read. Instead, it
created a new buffer for each handler in a pipeline. This works pretty
well as long as the number of handlers in a pipeline is only a few.
However, for a highly modular application with many handlers which
handles connections which lasts for relatively short period, it actually
makes the memory footprint issue much worse.
Changes:
All in all, this is about retaining all the good changes we made in 4 so
far such as better thread model and going back to the way how we dealt
with message events in 3.
To fix the memory consumption/footprint issue mentioned above, we made a
hard decision to break the backward compatibility again with the
following changes:
- Remove MessageBuf
- Merge Buf into ByteBuf
- Merge ChannelInboundByte/MessageHandler and ChannelStateHandler into ChannelInboundHandler
- Similar changes were made to the adapter classes
- Merge ChannelOutboundByte/MessageHandler and ChannelOperationHandler into ChannelOutboundHandler
- Similar changes were made to the adapter classes
- Introduce MessageList which is similar to `MessageEvent` in Netty 3
- Replace inboundBufferUpdated(ctx) with messageReceived(ctx, MessageList)
- Replace flush(ctx, promise) with write(ctx, MessageList, promise)
- Remove ByteToByteEncoder/Decoder/Codec
- Replaced by MessageToByteEncoder<ByteBuf>, ByteToMessageDecoder<ByteBuf>, and ByteMessageCodec<ByteBuf>
- Merge EmbeddedByteChannel and EmbeddedMessageChannel into EmbeddedChannel
- Add SimpleChannelInboundHandler which is sometimes more useful than
ChannelInboundHandlerAdapter
- Bring back Channel.isWritable() from Netty 3
- Add ChannelInboundHandler.channelWritabilityChanges() event
- Add RecvByteBufAllocator configuration property
- Similar to ReceiveBufferSizePredictor in Netty 3
- Some existing configuration properties such as
DatagramChannelConfig.receivePacketSize is gone now.
- Remove suspend/resumeIntermediaryDeallocation() in ByteBuf
This change would have been impossible without @normanmaurer's help. He
fixed, ported, and improved many parts of the changes.
2013-05-28 13:40:19 +02:00
|
|
|
public RxtxChannelConfig setWriteBufferHighWaterMark(int writeBufferHighWaterMark) {
|
2013-06-25 10:49:28 +02:00
|
|
|
super.setWriteBufferHighWaterMark(writeBufferHighWaterMark);
|
|
|
|
return this;
|
Revamp the core API to reduce memory footprint and consumption
The API changes made so far turned out to increase the memory footprint
and consumption while our intention was actually decreasing them.
Memory consumption issue:
When there are many connections which does not exchange data frequently,
the old Netty 4 API spent a lot more memory than 3 because it always
allocates per-handler buffer for each connection unless otherwise
explicitly stated by a user. In a usual real world load, a client
doesn't always send requests without pausing, so the idea of having a
buffer whose life cycle if bound to the life cycle of a connection
didn't work as expected.
Memory footprint issue:
The old Netty 4 API decreased overall memory footprint by a great deal
in many cases. It was mainly because the old Netty 4 API did not
allocate a new buffer and event object for each read. Instead, it
created a new buffer for each handler in a pipeline. This works pretty
well as long as the number of handlers in a pipeline is only a few.
However, for a highly modular application with many handlers which
handles connections which lasts for relatively short period, it actually
makes the memory footprint issue much worse.
Changes:
All in all, this is about retaining all the good changes we made in 4 so
far such as better thread model and going back to the way how we dealt
with message events in 3.
To fix the memory consumption/footprint issue mentioned above, we made a
hard decision to break the backward compatibility again with the
following changes:
- Remove MessageBuf
- Merge Buf into ByteBuf
- Merge ChannelInboundByte/MessageHandler and ChannelStateHandler into ChannelInboundHandler
- Similar changes were made to the adapter classes
- Merge ChannelOutboundByte/MessageHandler and ChannelOperationHandler into ChannelOutboundHandler
- Similar changes were made to the adapter classes
- Introduce MessageList which is similar to `MessageEvent` in Netty 3
- Replace inboundBufferUpdated(ctx) with messageReceived(ctx, MessageList)
- Replace flush(ctx, promise) with write(ctx, MessageList, promise)
- Remove ByteToByteEncoder/Decoder/Codec
- Replaced by MessageToByteEncoder<ByteBuf>, ByteToMessageDecoder<ByteBuf>, and ByteMessageCodec<ByteBuf>
- Merge EmbeddedByteChannel and EmbeddedMessageChannel into EmbeddedChannel
- Add SimpleChannelInboundHandler which is sometimes more useful than
ChannelInboundHandlerAdapter
- Bring back Channel.isWritable() from Netty 3
- Add ChannelInboundHandler.channelWritabilityChanges() event
- Add RecvByteBufAllocator configuration property
- Similar to ReceiveBufferSizePredictor in Netty 3
- Some existing configuration properties such as
DatagramChannelConfig.receivePacketSize is gone now.
- Remove suspend/resumeIntermediaryDeallocation() in ByteBuf
This change would have been impossible without @normanmaurer's help. He
fixed, ported, and improved many parts of the changes.
2013-05-28 13:40:19 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
@Override
|
|
|
|
public RxtxChannelConfig setWriteBufferLowWaterMark(int writeBufferLowWaterMark) {
|
2013-06-25 10:49:28 +02:00
|
|
|
super.setWriteBufferLowWaterMark(writeBufferLowWaterMark);
|
|
|
|
return this;
|
2013-03-08 08:20:46 +01:00
|
|
|
}
|
2013-01-20 11:18:40 +01:00
|
|
|
}
|