750 lines
27 KiB
Java
750 lines
27 KiB
Java
/*
|
|
* Copyright 2020 The Netty Project
|
|
*
|
|
* The Netty Project licenses this file to you under the Apache License,
|
|
* version 2.0 (the "License"); you may not use this file except in compliance
|
|
* with the License. You may obtain a copy of the License at:
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
* WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
* License for the specific language governing permissions and limitations
|
|
* under the License.
|
|
*/
|
|
package io.netty.channel.uring;
|
|
|
|
import io.netty.buffer.ByteBuf;
|
|
import io.netty.buffer.ByteBufAllocator;
|
|
import io.netty.buffer.ByteBufUtil;
|
|
import io.netty.buffer.Unpooled;
|
|
import io.netty.channel.AbstractChannel;
|
|
import io.netty.channel.Channel;
|
|
import io.netty.channel.ChannelConfig;
|
|
import io.netty.channel.ChannelFuture;
|
|
import io.netty.channel.ChannelFutureListener;
|
|
import io.netty.channel.ChannelMetadata;
|
|
import io.netty.channel.ChannelOutboundBuffer;
|
|
import io.netty.channel.ChannelPromise;
|
|
import io.netty.channel.ConnectTimeoutException;
|
|
import io.netty.channel.DefaultChannelConfig;
|
|
import io.netty.channel.EventLoop;
|
|
import io.netty.channel.RecvByteBufAllocator;
|
|
import io.netty.channel.socket.ChannelInputShutdownEvent;
|
|
import io.netty.channel.socket.ChannelInputShutdownReadComplete;
|
|
import io.netty.channel.socket.SocketChannelConfig;
|
|
import io.netty.channel.unix.Buffer;
|
|
import io.netty.channel.unix.Errors;
|
|
import io.netty.channel.unix.FileDescriptor;
|
|
import io.netty.channel.unix.IovArray;
|
|
import io.netty.channel.unix.NativeInetAddress;
|
|
import io.netty.channel.unix.Socket;
|
|
import io.netty.channel.unix.UnixChannel;
|
|
import io.netty.channel.unix.UnixChannelUtil;
|
|
import io.netty.util.ReferenceCountUtil;
|
|
import io.netty.util.internal.logging.InternalLogger;
|
|
import io.netty.util.internal.logging.InternalLoggerFactory;
|
|
|
|
import java.io.IOException;
|
|
import java.net.InetSocketAddress;
|
|
import java.net.SocketAddress;
|
|
import java.nio.ByteBuffer;
|
|
import java.nio.channels.AlreadyConnectedException;
|
|
import java.nio.channels.ClosedChannelException;
|
|
import java.nio.channels.ConnectionPendingException;
|
|
import java.nio.channels.NotYetConnectedException;
|
|
import java.nio.channels.UnresolvedAddressException;
|
|
import java.util.concurrent.ScheduledFuture;
|
|
import java.util.concurrent.TimeUnit;
|
|
|
|
import static io.netty.channel.unix.Errors.*;
|
|
import static io.netty.channel.unix.UnixChannelUtil.*;
|
|
import static io.netty.util.internal.ObjectUtil.*;
|
|
|
|
abstract class AbstractIOUringChannel extends AbstractChannel implements UnixChannel {
|
|
private static final InternalLogger logger = InternalLoggerFactory.getInstance(AbstractIOUringChannel.class);
|
|
private static final ChannelMetadata METADATA = new ChannelMetadata(false);
|
|
final LinuxSocket socket;
|
|
protected volatile boolean active;
|
|
|
|
// Different masks for outstanding I/O operations.
|
|
private static final int POLL_IN_SCHEDULED = 1;
|
|
private static final int POLL_OUT_SCHEDULED = 1 << 2;
|
|
private static final int WRITE_SCHEDULED = 1 << 3;
|
|
private static final int READ_SCHEDULED = 1 << 4;
|
|
private int ioState;
|
|
|
|
boolean inputClosedSeenErrorOnRead;
|
|
|
|
static final int SOCK_ADDR_LEN = 128;
|
|
|
|
/**
|
|
* The future of the current connection attempt. If not null, subsequent connection attempts will fail.
|
|
*/
|
|
private ChannelPromise connectPromise;
|
|
private ScheduledFuture<?> connectTimeoutFuture;
|
|
private SocketAddress requestedRemoteAddress;
|
|
private ByteBuffer remoteAddressMemory;
|
|
|
|
private volatile SocketAddress local;
|
|
private volatile SocketAddress remote;
|
|
|
|
AbstractIOUringChannel(final Channel parent, LinuxSocket socket) {
|
|
super(parent);
|
|
this.socket = checkNotNull(socket, "fd");
|
|
this.active = true;
|
|
|
|
if (active) {
|
|
// Directly cache the remote and local addresses
|
|
// See https://github.com/netty/netty/issues/2359
|
|
this.local = socket.localAddress();
|
|
this.remote = socket.remoteAddress();
|
|
}
|
|
|
|
if (parent != null) {
|
|
logger.trace("Create Channel Socket: {}", socket.intValue());
|
|
} else {
|
|
logger.trace("Create Server Socket: {}", socket.intValue());
|
|
}
|
|
}
|
|
|
|
AbstractIOUringChannel(final Channel parent, LinuxSocket socket, boolean active) {
|
|
super(parent);
|
|
this.socket = checkNotNull(socket, "fd");
|
|
this.active = active;
|
|
|
|
if (active) {
|
|
this.local = socket.localAddress();
|
|
this.remote = socket.remoteAddress();
|
|
}
|
|
|
|
if (parent != null) {
|
|
logger.trace("Create Channel Socket: {}", socket.intValue());
|
|
} else {
|
|
logger.trace("Create Server Socket: {}", socket.intValue());
|
|
}
|
|
}
|
|
|
|
AbstractIOUringChannel(Channel parent, LinuxSocket fd, SocketAddress remote) {
|
|
super(parent);
|
|
this.socket = checkNotNull(fd, "fd");
|
|
this.active = true;
|
|
|
|
// Directly cache the remote and local addresses
|
|
// See https://github.com/netty/netty/issues/2359
|
|
this.remote = remote;
|
|
this.local = fd.localAddress();
|
|
}
|
|
|
|
public boolean isOpen() {
|
|
return socket.isOpen();
|
|
}
|
|
|
|
@Override
|
|
public boolean isActive() {
|
|
return active;
|
|
}
|
|
|
|
@Override
|
|
public ChannelMetadata metadata() {
|
|
return METADATA;
|
|
}
|
|
|
|
@Override
|
|
public FileDescriptor fd() {
|
|
return socket;
|
|
}
|
|
|
|
@Override
|
|
protected abstract AbstractUringUnsafe newUnsafe();
|
|
|
|
@Override
|
|
protected boolean isCompatible(final EventLoop loop) {
|
|
return loop instanceof IOUringEventLoop;
|
|
}
|
|
|
|
protected final ByteBuf newDirectBuffer(ByteBuf buf) {
|
|
return newDirectBuffer(buf, buf);
|
|
}
|
|
|
|
protected final ByteBuf newDirectBuffer(Object holder, ByteBuf buf) {
|
|
final int readableBytes = buf.readableBytes();
|
|
if (readableBytes == 0) {
|
|
ReferenceCountUtil.release(holder);
|
|
return Unpooled.EMPTY_BUFFER;
|
|
}
|
|
|
|
final ByteBufAllocator alloc = alloc();
|
|
if (alloc.isDirectBufferPooled()) {
|
|
return newDirectBuffer0(holder, buf, alloc, readableBytes);
|
|
}
|
|
|
|
final ByteBuf directBuf = ByteBufUtil.threadLocalDirectBuffer();
|
|
if (directBuf == null) {
|
|
return newDirectBuffer0(holder, buf, alloc, readableBytes);
|
|
}
|
|
|
|
directBuf.writeBytes(buf, buf.readerIndex(), readableBytes);
|
|
ReferenceCountUtil.safeRelease(holder);
|
|
return directBuf;
|
|
}
|
|
|
|
private static ByteBuf newDirectBuffer0(Object holder, ByteBuf buf, ByteBufAllocator alloc, int capacity) {
|
|
final ByteBuf directBuf = alloc.directBuffer(capacity);
|
|
directBuf.writeBytes(buf, buf.readerIndex(), capacity);
|
|
ReferenceCountUtil.safeRelease(holder);
|
|
return directBuf;
|
|
}
|
|
|
|
@Override
|
|
protected void doDisconnect() throws Exception {
|
|
}
|
|
|
|
IOUringSubmissionQueue submissionQueue() {
|
|
IOUringEventLoop ioUringEventLoop = (IOUringEventLoop) eventLoop();
|
|
return ioUringEventLoop.getRingBuffer().getIoUringSubmissionQueue();
|
|
}
|
|
|
|
private void freeRemoteAddressMemory() {
|
|
if (remoteAddressMemory != null) {
|
|
Buffer.free(remoteAddressMemory);
|
|
remoteAddressMemory = null;
|
|
}
|
|
}
|
|
|
|
@Override
|
|
protected void doClose() throws Exception {
|
|
freeRemoteAddressMemory();
|
|
active = false;
|
|
|
|
// doClose() may be called by closeForcibly() before the Channel is registered on the EventLoop.
|
|
if (isRegistered()) {
|
|
IOUringSubmissionQueue submissionQueue = submissionQueue();
|
|
if ((ioState & POLL_IN_SCHEDULED) != 0) {
|
|
submissionQueue.addPollRemove(socket.intValue(), Native.POLLIN);
|
|
ioState &= ~POLL_IN_SCHEDULED;
|
|
}
|
|
if ((ioState & POLL_OUT_SCHEDULED) != 0) {
|
|
submissionQueue.addPollRemove(socket.intValue(), Native.POLLOUT);
|
|
ioState &= ~POLL_OUT_SCHEDULED;
|
|
}
|
|
submissionQueue.addPollRemove(socket.intValue(), Native.POLLRDHUP);
|
|
}
|
|
|
|
// Even if we allow half closed sockets we should give up on reading. Otherwise we may allow a read attempt on a
|
|
// socket which has not even been connected yet. This has been observed to block during unit tests.
|
|
//inputClosedSeenErrorOnRead = true;
|
|
try {
|
|
ChannelPromise promise = connectPromise;
|
|
if (promise != null) {
|
|
// Use tryFailure() instead of setFailure() to avoid the race against cancel().
|
|
promise.tryFailure(new ClosedChannelException());
|
|
connectPromise = null;
|
|
}
|
|
|
|
cancelConnectTimeoutFuture();
|
|
|
|
if (isRegistered()) {
|
|
// Need to check if we are on the EventLoop as doClose() may be triggered by the GlobalEventExecutor
|
|
// if SO_LINGER is used.
|
|
//
|
|
// See https://github.com/netty/netty/issues/7159
|
|
EventLoop loop = eventLoop();
|
|
if (loop.inEventLoop()) {
|
|
doDeregister();
|
|
} else {
|
|
loop.execute(new Runnable() {
|
|
@Override
|
|
public void run() {
|
|
try {
|
|
doDeregister();
|
|
} catch (Throwable cause) {
|
|
pipeline().fireExceptionCaught(cause);
|
|
}
|
|
}
|
|
});
|
|
}
|
|
}
|
|
} finally {
|
|
socket.close();
|
|
}
|
|
}
|
|
|
|
//deregister
|
|
// Channel/ChannelHandlerContext.read() was called
|
|
@Override
|
|
protected void doBeginRead() {
|
|
final AbstractUringUnsafe unsafe = (AbstractUringUnsafe) unsafe();
|
|
if ((ioState & POLL_IN_SCHEDULED) == 0) {
|
|
unsafe.schedulePollIn();
|
|
}
|
|
}
|
|
|
|
@Override
|
|
protected void doWrite(ChannelOutboundBuffer in) {
|
|
if ((ioState & WRITE_SCHEDULED) != 0) {
|
|
return;
|
|
}
|
|
|
|
int msgCount = in.size();
|
|
if (msgCount == 0) {
|
|
return;
|
|
}
|
|
ByteBuf msg = (ByteBuf) in.current();
|
|
if (msgCount > 1 ||
|
|
// We also need some special handling for CompositeByteBuf
|
|
msg.nioBufferCount() > 1) {
|
|
doWriteMultiple(in);
|
|
} else if (msgCount == 1) {
|
|
doWriteSingle(msg);
|
|
}
|
|
}
|
|
|
|
private void doWriteMultiple(ChannelOutboundBuffer in) {
|
|
|
|
final IovArray iovecArray = ((IOUringEventLoop) eventLoop()).getIovArray();
|
|
int offset = iovecArray.count();
|
|
try {
|
|
in.forEachFlushedMessage(iovecArray);
|
|
|
|
int count = iovecArray.count() - offset;
|
|
if (count > 0) {
|
|
submissionQueue().addWritev(socket.intValue(), iovecArray.memoryAddress(offset), count);
|
|
submissionQueue().submit();
|
|
|
|
ioState |= WRITE_SCHEDULED;
|
|
if (iovecArray.isFull()) {
|
|
submissionQueue().submit();
|
|
iovecArray.clear();
|
|
}
|
|
} else {
|
|
in.removeBytes(0);
|
|
}
|
|
} catch (Exception e) {
|
|
e.printStackTrace();
|
|
// This should never happem, anyway fallback to single write.
|
|
doWriteSingle((ByteBuf) in.current());
|
|
}
|
|
|
|
}
|
|
|
|
protected final void doWriteSingle(ByteBuf buf) {
|
|
IOUringSubmissionQueue submissionQueue = submissionQueue();
|
|
submissionQueue.addWrite(socket.intValue(), buf.memoryAddress(), buf.readerIndex(),
|
|
buf.writerIndex());
|
|
ioState |= WRITE_SCHEDULED;
|
|
}
|
|
|
|
//POLLOUT
|
|
private void addPollOut() {
|
|
assert (ioState & POLL_OUT_SCHEDULED) == 0;
|
|
ioState |= POLL_OUT_SCHEDULED;
|
|
IOUringSubmissionQueue submissionQueue = submissionQueue();
|
|
submissionQueue.addPollOut(socket.intValue());
|
|
}
|
|
|
|
abstract class AbstractUringUnsafe extends AbstractUnsafe {
|
|
private IOUringRecvByteAllocatorHandle allocHandle;
|
|
|
|
@Override
|
|
protected final void flush0() {
|
|
// Flush immediately only when there's no pending flush.
|
|
// If there's a pending flush operation, event loop will call forceFlush() later,
|
|
// and thus there's no need to call it now.
|
|
if ((ioState & POLL_OUT_SCHEDULED) == 0) {
|
|
super.flush0();
|
|
}
|
|
}
|
|
|
|
private void fulfillConnectPromise(ChannelPromise promise, Throwable cause) {
|
|
if (promise == null) {
|
|
// Closed via cancellation and the promise has been notified already.
|
|
return;
|
|
}
|
|
|
|
// Use tryFailure() instead of setFailure() to avoid the race against cancel().
|
|
promise.tryFailure(cause);
|
|
closeIfClosed();
|
|
}
|
|
|
|
private void fulfillConnectPromise(ChannelPromise promise, boolean wasActive) {
|
|
if (promise == null) {
|
|
// Closed via cancellation and the promise has been notified already.
|
|
return;
|
|
}
|
|
active = true;
|
|
|
|
if (local == null) {
|
|
local = socket.localAddress();
|
|
}
|
|
computeRemote();
|
|
|
|
// Register POLLRDHUP
|
|
IOUringSubmissionQueue submissionQueue = submissionQueue();
|
|
submissionQueue.addPollRdHup(fd().intValue());
|
|
|
|
// Get the state as trySuccess() may trigger an ChannelFutureListener that will close the Channel.
|
|
// We still need to ensure we call fireChannelActive() in this case.
|
|
boolean active = isActive();
|
|
|
|
// trySuccess() will return false if a user cancelled the connection attempt.
|
|
boolean promiseSet = promise.trySuccess();
|
|
|
|
// Regardless if the connection attempt was cancelled, channelActive() event should be triggered,
|
|
// because what happened is what happened.
|
|
if (!wasActive && active) {
|
|
pipeline().fireChannelActive();
|
|
}
|
|
|
|
// If a user cancelled the connection attempt, close the channel, which is followed by channelInactive().
|
|
if (!promiseSet) {
|
|
close(voidPromise());
|
|
}
|
|
}
|
|
|
|
IOUringRecvByteAllocatorHandle newIOUringHandle(RecvByteBufAllocator.ExtendedHandle handle) {
|
|
return new IOUringRecvByteAllocatorHandle(handle);
|
|
}
|
|
|
|
@Override
|
|
public IOUringRecvByteAllocatorHandle recvBufAllocHandle() {
|
|
if (allocHandle == null) {
|
|
allocHandle = newIOUringHandle((RecvByteBufAllocator.ExtendedHandle) super.recvBufAllocHandle());
|
|
}
|
|
return allocHandle;
|
|
}
|
|
|
|
void shutdownInput(boolean rdHup) {
|
|
logger.trace("shutdownInput Fd: {}", fd().intValue());
|
|
if (!socket.isInputShutdown()) {
|
|
if (isAllowHalfClosure(config())) {
|
|
try {
|
|
socket.shutdown(true, false);
|
|
} catch (IOException ignored) {
|
|
// We attempted to shutdown and failed, which means the input has already effectively been
|
|
// shutdown.
|
|
fireEventAndClose(ChannelInputShutdownEvent.INSTANCE);
|
|
return;
|
|
} catch (NotYetConnectedException ignore) {
|
|
// We attempted to shutdown and failed, which means the input has already effectively been
|
|
// shutdown.
|
|
}
|
|
pipeline().fireUserEventTriggered(ChannelInputShutdownEvent.INSTANCE);
|
|
} else {
|
|
close(voidPromise());
|
|
}
|
|
} else if (!rdHup) {
|
|
inputClosedSeenErrorOnRead = true;
|
|
pipeline().fireUserEventTriggered(ChannelInputShutdownReadComplete.INSTANCE);
|
|
}
|
|
}
|
|
|
|
private void fireEventAndClose(Object evt) {
|
|
pipeline().fireUserEventTriggered(evt);
|
|
close(voidPromise());
|
|
}
|
|
|
|
void schedulePollIn() {
|
|
assert (ioState & POLL_IN_SCHEDULED) == 0;
|
|
if (!isActive() || shouldBreakIoUringInReady(config())) {
|
|
return;
|
|
}
|
|
ioState |= POLL_IN_SCHEDULED;
|
|
IOUringSubmissionQueue submissionQueue = submissionQueue();
|
|
submissionQueue.addPollIn(socket.intValue());
|
|
}
|
|
|
|
final void readComplete(int res) {
|
|
ioState &= ~READ_SCHEDULED;
|
|
|
|
readComplete0(res);
|
|
}
|
|
|
|
protected abstract void readComplete0(int res);
|
|
|
|
/**
|
|
* Called once POLLRDHUP event is ready to be processed
|
|
*/
|
|
final void pollRdHup(int res) {
|
|
if (isActive()) {
|
|
if ((ioState & READ_SCHEDULED) == 0) {
|
|
scheduleFirstRead();
|
|
}
|
|
} else {
|
|
// Just to be safe make sure the input marked as closed.
|
|
shutdownInput(true);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Called once POLLIN event is ready to be processed
|
|
*/
|
|
final void pollIn(int res) {
|
|
ioState &= ~POLL_IN_SCHEDULED;
|
|
|
|
if ((ioState & READ_SCHEDULED) == 0) {
|
|
scheduleFirstRead();
|
|
}
|
|
}
|
|
|
|
private void scheduleFirstRead() {
|
|
// This is a new "read loop" so we need to reset the allocHandle.
|
|
final ChannelConfig config = config();
|
|
final IOUringRecvByteAllocatorHandle allocHandle = recvBufAllocHandle();
|
|
allocHandle.reset(config);
|
|
scheduleRead();
|
|
}
|
|
|
|
protected final void scheduleRead() {
|
|
ioState |= READ_SCHEDULED;
|
|
scheduleRead0();
|
|
}
|
|
|
|
protected abstract void scheduleRead0();
|
|
|
|
/**
|
|
* Called once POLLOUT event is ready to be processed
|
|
*/
|
|
final void pollOut(int res) {
|
|
ioState &= ~POLL_OUT_SCHEDULED;
|
|
|
|
// pending connect
|
|
if (connectPromise != null) {
|
|
// Note this method is invoked by the event loop only if the connection attempt was
|
|
// neither cancelled nor timed out.
|
|
|
|
assert eventLoop().inEventLoop();
|
|
|
|
boolean connectStillInProgress = false;
|
|
try {
|
|
boolean wasActive = isActive();
|
|
if (!socket.finishConnect()) {
|
|
connectStillInProgress = true;
|
|
return;
|
|
}
|
|
fulfillConnectPromise(connectPromise, wasActive);
|
|
} catch (Throwable t) {
|
|
fulfillConnectPromise(connectPromise, annotateConnectException(t, requestedRemoteAddress));
|
|
} finally {
|
|
if (!connectStillInProgress) {
|
|
// Check for null as the connectTimeoutFuture is only created if a connectTimeoutMillis > 0
|
|
// is used
|
|
// See https://github.com/netty/netty/issues/1770
|
|
cancelConnectTimeoutFuture();
|
|
connectPromise = null;
|
|
} else {
|
|
// The connect was not done yet, register for POLLOUT again
|
|
addPollOut();
|
|
}
|
|
}
|
|
} else if (!getSocket().isOutputShutdown()) {
|
|
// Try writing again
|
|
super.flush0();
|
|
}
|
|
}
|
|
|
|
final void writeComplete(int res) {
|
|
ioState &= ~WRITE_SCHEDULED;
|
|
|
|
ChannelOutboundBuffer channelOutboundBuffer = unsafe().outboundBuffer();
|
|
if (res >= 0) {
|
|
channelOutboundBuffer.removeBytes(res);
|
|
doWrite(channelOutboundBuffer);
|
|
} else {
|
|
try {
|
|
if (ioResult("io_uring write", res) == 0) {
|
|
// We were not able to write everything, let's register for POLLOUT
|
|
addPollOut();
|
|
}
|
|
} catch (Throwable cause) {
|
|
handleWriteError(cause);
|
|
}
|
|
}
|
|
}
|
|
|
|
final void connectComplete(int res) {
|
|
freeRemoteAddressMemory();
|
|
|
|
if (res == 0) {
|
|
fulfillConnectPromise(connectPromise, active);
|
|
} else {
|
|
if (res == ERRNO_EINPROGRESS_NEGATIVE) {
|
|
// connect not complete yet need to wait for poll_out event
|
|
addPollOut();
|
|
} else {
|
|
try {
|
|
Errors.throwConnectException("io_uring connect", res);
|
|
} catch (Throwable cause) {
|
|
fulfillConnectPromise(connectPromise, cause);
|
|
} finally {
|
|
// Check for null as the connectTimeoutFuture is only created if a connectTimeoutMillis > 0 is
|
|
// used
|
|
// See https://github.com/netty/netty/issues/1770
|
|
cancelConnectTimeoutFuture();
|
|
connectPromise = null;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
@Override
|
|
public void connect(
|
|
final SocketAddress remoteAddress, final SocketAddress localAddress, final ChannelPromise promise) {
|
|
if (!promise.setUncancellable() || !ensureOpen(promise)) {
|
|
return;
|
|
}
|
|
|
|
try {
|
|
if (connectPromise != null) {
|
|
throw new ConnectionPendingException();
|
|
}
|
|
|
|
doConnect(remoteAddress, localAddress);
|
|
InetSocketAddress inetSocketAddress = (InetSocketAddress) remoteAddress;
|
|
NativeInetAddress address = NativeInetAddress.newInstance(inetSocketAddress.getAddress());
|
|
|
|
remoteAddressMemory = Buffer.allocateDirectWithNativeOrder(SOCK_ADDR_LEN);
|
|
long remoteAddressMemoryAddress = Buffer.memoryAddress(remoteAddressMemory);
|
|
|
|
socket.initAddress(address.address(), address.scopeId(), inetSocketAddress.getPort(),
|
|
remoteAddressMemoryAddress);
|
|
final IOUringSubmissionQueue ioUringSubmissionQueue = submissionQueue();
|
|
ioUringSubmissionQueue.addConnect(socket.intValue(), remoteAddressMemoryAddress, SOCK_ADDR_LEN);
|
|
} catch (Throwable t) {
|
|
closeIfClosed();
|
|
promise.tryFailure(annotateConnectException(t, remoteAddress));
|
|
return;
|
|
}
|
|
connectPromise = promise;
|
|
requestedRemoteAddress = remoteAddress;
|
|
// Schedule connect timeout.
|
|
int connectTimeoutMillis = config().getConnectTimeoutMillis();
|
|
if (connectTimeoutMillis > 0) {
|
|
connectTimeoutFuture = eventLoop().schedule(new Runnable() {
|
|
@Override
|
|
public void run() {
|
|
ChannelPromise connectPromise = AbstractIOUringChannel.this.connectPromise;
|
|
ConnectTimeoutException cause =
|
|
new ConnectTimeoutException("connection timed out: " + remoteAddress);
|
|
if (connectPromise != null && connectPromise.tryFailure(cause)) {
|
|
close(voidPromise());
|
|
}
|
|
}
|
|
}, connectTimeoutMillis, TimeUnit.MILLISECONDS);
|
|
}
|
|
|
|
promise.addListener(new ChannelFutureListener() {
|
|
@Override
|
|
public void operationComplete(ChannelFuture future) {
|
|
if (future.isCancelled()) {
|
|
cancelConnectTimeoutFuture();
|
|
connectPromise = null;
|
|
close(voidPromise());
|
|
}
|
|
}
|
|
});
|
|
}
|
|
}
|
|
|
|
@Override
|
|
protected Object filterOutboundMessage(Object msg) {
|
|
if (msg instanceof ByteBuf) {
|
|
ByteBuf buf = (ByteBuf) msg;
|
|
return UnixChannelUtil.isBufferCopyNeededForWrite(buf)? newDirectBuffer(buf) : buf;
|
|
}
|
|
|
|
throw new UnsupportedOperationException("unsupported message type");
|
|
}
|
|
|
|
@Override
|
|
protected void doRegister() throws Exception {
|
|
((IOUringEventLoop) eventLoop()).add(this);
|
|
}
|
|
|
|
@Override
|
|
protected void doDeregister() throws Exception {
|
|
((IOUringEventLoop) eventLoop()).remove(this);
|
|
}
|
|
|
|
@Override
|
|
public void doBind(final SocketAddress local) throws Exception {
|
|
if (local instanceof InetSocketAddress) {
|
|
checkResolvable((InetSocketAddress) local);
|
|
}
|
|
socket.bind(local);
|
|
this.local = socket.localAddress();
|
|
}
|
|
|
|
protected static void checkResolvable(InetSocketAddress addr) {
|
|
if (addr.isUnresolved()) {
|
|
throw new UnresolvedAddressException();
|
|
}
|
|
}
|
|
|
|
@Override
|
|
public abstract DefaultChannelConfig config();
|
|
|
|
@Override
|
|
protected SocketAddress localAddress0() {
|
|
return local;
|
|
}
|
|
|
|
@Override
|
|
protected SocketAddress remoteAddress0() {
|
|
return remote;
|
|
}
|
|
|
|
protected Socket getSocket() {
|
|
return socket;
|
|
}
|
|
|
|
/**
|
|
* Connect to the remote peer
|
|
*/
|
|
private void doConnect(SocketAddress remoteAddress, SocketAddress localAddress) throws Exception {
|
|
if (localAddress instanceof InetSocketAddress) {
|
|
checkResolvable((InetSocketAddress) localAddress);
|
|
}
|
|
|
|
InetSocketAddress remoteSocketAddr = remoteAddress instanceof InetSocketAddress
|
|
? (InetSocketAddress) remoteAddress : null;
|
|
if (remoteSocketAddr != null) {
|
|
checkResolvable(remoteSocketAddr);
|
|
}
|
|
|
|
if (remote != null) {
|
|
// Check if already connected before trying to connect. This is needed as connect(...) will not return -1
|
|
// and set errno to EISCONN if a previous connect(...) attempt was setting errno to EINPROGRESS and finished
|
|
// later.
|
|
throw new AlreadyConnectedException();
|
|
}
|
|
|
|
if (localAddress != null) {
|
|
socket.bind(localAddress);
|
|
}
|
|
}
|
|
|
|
private static boolean isAllowHalfClosure(ChannelConfig config) {
|
|
return config instanceof SocketChannelConfig &&
|
|
((SocketChannelConfig) config).isAllowHalfClosure();
|
|
}
|
|
|
|
private void cancelConnectTimeoutFuture() {
|
|
if (connectTimeoutFuture != null) {
|
|
connectTimeoutFuture.cancel(false);
|
|
connectTimeoutFuture = null;
|
|
}
|
|
}
|
|
|
|
private void computeRemote() {
|
|
if (requestedRemoteAddress instanceof InetSocketAddress) {
|
|
remote = computeRemoteAddr((InetSocketAddress) requestedRemoteAddress, socket.remoteAddress());
|
|
}
|
|
}
|
|
|
|
private boolean shouldBreakIoUringInReady(ChannelConfig config) {
|
|
return socket.isInputShutdown() && (inputClosedSeenErrorOnRead || !isAllowHalfClosure(config));
|
|
}
|
|
}
|