2 * Socket and pipe I/O utilities used in rsync.
4 * Copyright (C) 1996-2001 Andrew Tridgell
5 * Copyright (C) 1996 Paul Mackerras
6 * Copyright (C) 2001, 2002 Martin Pool <mbp@samba.org>
7 * Copyright (C) 2003-2009 Wayne Davison
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License as published by
11 * the Free Software Foundation; either version 3 of the License, or
12 * (at your option) any later version.
14 * This program is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 * GNU General Public License for more details.
19 * You should have received a copy of the GNU General Public License along
20 * with this program; if not, visit the http://fsf.org website.
23 /* Rsync provides its own multiplexing system, which is used to send
24 * stderr and stdout over a single socket.
26 * For historical reasons this is off during the start of the
27 * connection, but it's switched on quite early using
28 * io_start_multiplex_out() and io_start_multiplex_in(). */
34 /** If no timeout is specified then use a 60 second select timeout */
35 #define SELECT_TIMEOUT 60
38 extern size_t bwlimit_writemax
;
39 extern int io_timeout
;
42 extern int am_receiver
;
43 extern int am_generator
;
44 extern int msgs2stderr
;
45 extern int inc_recurse
;
49 extern int file_total
;
50 extern int file_old_total
;
52 extern int read_batch
;
53 extern int compat_flags
;
54 extern int protect_args
;
55 extern int checksum_seed
;
56 extern int protocol_version
;
57 extern int remove_source_files
;
58 extern int preserve_hard_links
;
59 extern BOOL extra_flist_sending_enabled
;
60 extern struct stats stats
;
61 extern struct file_list
*cur_flist
;
63 extern int filesfrom_convert
;
64 extern iconv_t ic_send
, ic_recv
;
67 int csum_length
= SHORT_SUM_LENGTH
; /* initial value */
71 int forward_flist_data
= 0;
72 BOOL flist_receiving_enabled
= False
;
74 /* Ignore an EOF error if non-zero. See whine_about_eof(). */
75 int kluge_around_eof
= 0;
80 int64 total_data_read
= 0;
81 int64 total_data_written
= 0;
86 int out_fd
; /* Both "out" and "msg" go to this fd. */
88 unsigned out_empty_len
;
89 size_t raw_data_header_pos
; /* in the out xbuf */
90 size_t raw_flushing_ends_before
; /* in the out xbuf */
91 size_t raw_input_ends_before
; /* in the in xbuf */
92 } iobuf
= { .in_fd
= -1, .out_fd
= -1 };
94 static time_t last_io_in
;
95 static time_t last_io_out
;
97 static int write_batch_monitor_in
= -1;
98 static int write_batch_monitor_out
= -1;
100 static int ff_forward_fd
= -1;
101 static int ff_reenable_multiplex
= -1;
102 static char ff_lastchar
= '\0';
103 static xbuf ff_xb
= EMPTY_XBUF
;
105 static xbuf iconv_buf
= EMPTY_XBUF
;
107 static int select_timeout
= SELECT_TIMEOUT
;
108 static int active_filecnt
= 0;
109 static OFF_T active_bytecnt
= 0;
110 static int first_message
= 1;
112 static char int_byte_extra
[64] = {
113 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, /* (00 - 3F)/4 */
114 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, /* (40 - 7F)/4 */
115 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, /* (80 - BF)/4 */
116 2, 2, 2, 2, 2, 2, 2, 2, 3, 3, 3, 3, 4, 4, 5, 6, /* (C0 - FF)/4 */
119 /* Our I/O buffers are sized with no bits on in the lowest byte of the "size"
120 * (indeed, our rounding of sizes in 1024-byte units assures more than this).
121 * This allows the code that is storing bytes near the physical end of a
122 * circular buffer to temporarily reduce the buffer's size (in order to make
123 * some storing idioms easier), while also making it simple to restore the
124 * buffer's actual size when the buffer's "pos" wraps around to the start (we
125 * just round the buffer's size up again). */
127 #define IOBUF_WAS_REDUCED(siz) ((siz) & 0xFF)
128 #define IOBUF_RESTORE_SIZE(siz) (((siz) | 0xFF) + 1)
130 #define IN_MULTIPLEXED (iobuf.in_multiplexed != 0)
131 #define IN_MULTIPLEXED_AND_READY (iobuf.in_multiplexed > 0)
132 #define OUT_MULTIPLEXED (iobuf.out_empty_len != 0)
134 #define PIO_NEED_INPUT (1<<0) /* The *_NEED_* flags are mutually exclusive. */
135 #define PIO_NEED_OUTROOM (1<<1)
136 #define PIO_NEED_MSGROOM (1<<2)
138 #define PIO_CONSUME_INPUT (1<<4) /* Must becombined with PIO_NEED_INPUT. */
140 #define PIO_INPUT_AND_CONSUME (PIO_NEED_INPUT | PIO_CONSUME_INPUT)
141 #define PIO_NEED_FLAGS (PIO_NEED_INPUT | PIO_NEED_OUTROOM | PIO_NEED_MSGROOM)
143 #define REMOTE_OPTION_ERROR "rsync: on remote machine: -"
144 #define REMOTE_OPTION_ERROR2 ": unknown option"
146 #define FILESFROM_BUFLEN 2048
148 enum festatus
{ FES_SUCCESS
, FES_REDO
, FES_NO_SEND
};
150 static flist_ndx_list redo_list
, hlink_list
;
152 static void read_a_msg(void);
153 static void drain_multiplex_messages(void);
154 static void sleep_for_bwlimit(int bytes_written
);
156 static void check_timeout(BOOL allow_keepalive
)
160 /* On the receiving side, the generator is now the one that decides
161 * when a timeout has occurred. When it is sifting through a lot of
162 * files looking for work, it will be sending keep-alive messages to
163 * the sender, and even though the receiver won't be sending/receiving
164 * anything (not even keep-alive messages), the successful writes to
165 * the sender will keep things going. If the receiver is actively
166 * receiving data, it will ensure that the generator knows that it is
167 * not idle by sending the generator keep-alive messages (since the
168 * generator might be blocked trying to send checksums, it needs to
169 * know that the receiver is active). Thus, as long as one or the
170 * other is successfully doing work, the generator will not timeout. */
176 if (allow_keepalive
) {
177 /* This may put data into iobuf.msg w/o flushing. */
178 maybe_send_keepalive(t
, 0);
187 chk
= MAX(last_io_out
, last_io_in
);
188 if (t
- chk
>= io_timeout
) {
191 rprintf(FERROR
, "[%s] io timeout after %d seconds -- exiting\n",
192 who_am_i(), (int)(t
-chk
));
193 exit_cleanup(RERR_TIMEOUT
);
197 /* It's almost always an error to get an EOF when we're trying to read from the
198 * network, because the protocol is (for the most part) self-terminating.
200 * There is one case for the receiver when it is at the end of the transfer
201 * (hanging around reading any keep-alive packets that might come its way): if
202 * the sender dies before the generator's kill-signal comes through, we can end
203 * up here needing to loop until the kill-signal arrives. In this situation,
204 * kluge_around_eof will be < 0.
206 * There is another case for older protocol versions (< 24) where the module
207 * listing was not terminated, so we must ignore an EOF error in that case and
208 * exit. In this situation, kluge_around_eof will be > 0. */
209 static NORETURN
void whine_about_eof(BOOL allow_kluge
)
211 if (kluge_around_eof
&& allow_kluge
) {
213 if (kluge_around_eof
> 0)
215 /* If we're still here after 10 seconds, exit with an error. */
216 for (i
= 10*1000/20; i
--; )
220 rprintf(FERROR
, RSYNC_NAME
": connection unexpectedly closed "
221 "(%s bytes received so far) [%s]\n",
222 big_num(stats
.total_read
), who_am_i());
224 exit_cleanup(RERR_STREAMIO
);
227 /* Do a safe read, handling any needed looping and error handling.
228 * Returns the count of the bytes read, which will only be different
229 * from "len" if we encountered an EOF. This routine is not used on
230 * the socket except very early in the transfer. */
231 static size_t safe_read(int fd
, char *buf
, size_t len
)
236 assert(fd
!= iobuf
.in_fd
);
238 n
= read(fd
, buf
, len
);
239 if ((size_t)n
== len
|| n
== 0) {
240 if (DEBUG_GTE(IO
, 2))
241 rprintf(FINFO
, "[%s] safe_read(%d)=%ld\n", who_am_i(), fd
, (long)n
);
245 if (errno
!= EINTR
&& errno
!= EWOULDBLOCK
&& errno
!= EAGAIN
) {
247 rsyserr(FERROR
, errno
, "safe_read failed to read %ld bytes [%s]",
248 (long)len
, who_am_i());
249 exit_cleanup(RERR_STREAMIO
);
264 tv
.tv_sec
= select_timeout
;
267 cnt
= select(fd
+1, &r_fds
, NULL
, &e_fds
, &tv
);
269 if (cnt
< 0 && errno
== EBADF
) {
270 rsyserr(FERROR
, errno
, "safe_read select failed [%s]",
272 exit_cleanup(RERR_FILEIO
);
275 maybe_send_keepalive(time(NULL
), MSK_ALLOW_FLUSH
);
279 /*if (FD_ISSET(fd, &e_fds))
280 rprintf(FINFO, "select exception on fd %d\n", fd); */
282 if (FD_ISSET(fd
, &r_fds
)) {
283 n
= read(fd
, buf
+ got
, len
- got
);
284 if (DEBUG_GTE(IO
, 2))
285 rprintf(FINFO
, "[%s] safe_read(%d)=%ld\n", who_am_i(), fd
, (long)n
);
293 if ((got
+= (size_t)n
) == len
)
301 static const char *what_fd_is(int fd
)
305 if (fd
== sock_f_out
)
307 else if (fd
== iobuf
.out_fd
)
309 else if (fd
== batch_fd
)
312 snprintf(buf
, sizeof buf
, "fd %d", fd
);
317 /* Do a safe write, handling any needed looping and error handling.
318 * Returns only if everything was successfully written. This routine
319 * is not used on the socket except very early in the transfer. */
320 static void safe_write(int fd
, const char *buf
, size_t len
)
324 assert(fd
!= iobuf
.out_fd
);
326 n
= write(fd
, buf
, len
);
327 if ((size_t)n
== len
)
330 if (errno
!= EINTR
&& errno
!= EWOULDBLOCK
&& errno
!= EAGAIN
) {
332 rsyserr(FERROR
, errno
,
333 "safe_write failed to write %ld bytes to %s [%s]",
334 (long)len
, what_fd_is(fd
), who_am_i());
335 exit_cleanup(RERR_STREAMIO
);
349 tv
.tv_sec
= select_timeout
;
352 cnt
= select(fd
+ 1, NULL
, &w_fds
, NULL
, &tv
);
354 if (cnt
< 0 && errno
== EBADF
) {
355 rsyserr(FERROR
, errno
, "safe_write select failed on %s [%s]",
356 what_fd_is(fd
), who_am_i());
357 exit_cleanup(RERR_FILEIO
);
360 maybe_send_keepalive(time(NULL
), MSK_ALLOW_FLUSH
);
364 if (FD_ISSET(fd
, &w_fds
)) {
365 n
= write(fd
, buf
, len
);
377 /* This is only called when files-from data is known to be available. We read
378 * a chunk of data and put it into the output buffer. */
379 static void forward_filesfrom_data(void)
383 len
= read(ff_forward_fd
, ff_xb
.buf
+ ff_xb
.len
, ff_xb
.size
- ff_xb
.len
);
385 if (len
== 0 || errno
!= EINTR
) {
386 /* Send end-of-file marker */
388 write_buf(iobuf
.out_fd
, "\0\0", ff_lastchar
? 2 : 1);
390 if (ff_reenable_multiplex
>= 0)
391 io_start_multiplex_out(ff_reenable_multiplex
);
396 if (DEBUG_GTE(IO
, 2))
397 rprintf(FINFO
, "[%s] files-from read=%ld\n", who_am_i(), (long)len
);
404 char *s
= ff_xb
.buf
+ len
;
405 /* Transform CR and/or LF into '\0' */
406 while (s
-- > ff_xb
.buf
) {
407 if (*s
== '\n' || *s
== '\r')
416 /* Last buf ended with a '\0', so don't let this buf start with one. */
417 while (len
&& *s
== '\0')
419 ff_xb
.pos
= s
- ff_xb
.buf
;
423 if (filesfrom_convert
&& len
) {
424 char *sob
= ff_xb
.buf
+ ff_xb
.pos
, *s
= sob
;
425 char *eob
= sob
+ len
;
426 int flags
= ICB_INCLUDE_BAD
| ICB_INCLUDE_INCOMPLETE
| ICB_CIRCULAR_OUT
;
427 if (ff_lastchar
== '\0')
429 /* Convert/send each null-terminated string separately, skipping empties. */
432 ff_xb
.len
= s
- sob
- 1;
433 if (iconvbufs(ic_send
, &ff_xb
, &iobuf
.out
, flags
) < 0)
434 exit_cleanup(RERR_PROTOCOL
); /* impossible? */
435 write_buf(iobuf
.out_fd
, s
-1, 1); /* Send the '\0'. */
436 while (s
!= eob
&& *s
== '\0')
439 ff_xb
.pos
= sob
- ff_xb
.buf
;
444 if ((ff_xb
.len
= s
- sob
) == 0)
447 /* Handle a partial string specially, saving any incomplete chars. */
448 flags
&= ~ICB_INCLUDE_INCOMPLETE
;
449 if (iconvbufs(ic_send
, &ff_xb
, &iobuf
.out
, flags
) < 0) {
451 exit_cleanup(RERR_PROTOCOL
); /* impossible? */
453 memmove(ff_xb
.buf
, ff_xb
.buf
+ ff_xb
.pos
, ff_xb
.len
);
455 ff_lastchar
= 'x'; /* Anything non-zero. */
461 char *f
= ff_xb
.buf
+ ff_xb
.pos
;
464 /* Eliminate any multi-'\0' runs. */
466 if (!(*t
++ = *f
++)) {
467 while (f
!= eob
&& *f
== '\0')
472 if ((len
= t
- ff_xb
.buf
) != 0) {
473 /* This will not circle back to perform_io() because we only get
474 * called when there is plenty of room in the output buffer. */
475 write_buf(iobuf
.out_fd
, ff_xb
.buf
, len
);
480 void reduce_iobuf_size(xbuf
*out
, size_t new_size
)
482 if (new_size
< out
->size
) {
483 if (DEBUG_GTE(IO
, 4)) {
484 const char *name
= out
== &iobuf
.out
? "iobuf.out"
485 : out
== &iobuf
.msg
? "iobuf.msg"
488 rprintf(FINFO
, "[%s] reduced size of %s (-%d)\n",
489 who_am_i(), name
, (int)(out
->size
- new_size
));
492 out
->size
= new_size
;
496 void restore_iobuf_size(xbuf
*out
)
498 if (IOBUF_WAS_REDUCED(out
->size
)) {
499 size_t new_size
= IOBUF_RESTORE_SIZE(out
->size
);
500 if (DEBUG_GTE(IO
, 4)) {
501 const char *name
= out
== &iobuf
.out
? "iobuf.out"
502 : out
== &iobuf
.msg
? "iobuf.msg"
505 rprintf(FINFO
, "[%s] restored size of %s (+%d)\n",
506 who_am_i(), name
, (int)(new_size
- out
->size
));
509 out
->size
= new_size
;
513 /* Perform buffered input and/or output until specified conditions are met.
514 * When given a "needed" read or write request, this returns without doing any
515 * I/O if the needed input bytes or write space is already available. Once I/O
516 * is needed, this will try to do whatever reading and/or writing is currently
517 * possible, up to the maximum buffer allowances, no matter if this is a read
518 * or write request. However, the I/O stops as soon as the required input
519 * bytes or output space is available. If this is not a read request, the
520 * routine may also do some advantageous reading of messages from a multiplexed
521 * input source (which ensures that we don't jam up with everyone in their
522 * "need to write" code and nobody reading the accumulated data that would make
525 * The iobuf.in, .out and .msg buffers are all circular. Callers need to be
526 * aware that some data copies will need to be split when the bytes wrap around
527 * from the end to the start. In order to help make writing into the output
528 * buffers easier for some operations (such as the use of SIVAL() into the
529 * buffer) a buffer may be temporarily shortened by a small amount, but the
530 * original size will be automatically restored when the .pos wraps to the
531 * start. See also the 3 raw_* iobuf vars that are used in the handling of
532 * MSG_DATA bytes as they are read-from/written-into the buffers.
534 * When writing, we flush data in the following priority order:
536 * 1. Finish writing any in-progress MSG_DATA sequence from iobuf.out.
538 * 2. Write out all the messages from the message buf (if iobuf.msg is active).
539 * Yes, this means that a PIO_NEED_OUTROOM call will completely flush any
540 * messages before getting to the iobuf.out flushing (except for rule 1).
542 * 3. Write out the raw data from iobuf.out, possibly filling in the multiplexed
543 * MSG_DATA header that was pre-allocated (when output is multiplexed).
545 * TODO: items for possible future work:
547 * - Make this routine able to read the generator-to-receiver batch flow?
549 * Unlike the old routines that this replaces, it is OK to read ahead as far as
550 * we can because the read_a_msg() routine now reads its bytes out of the input
551 * buffer. In the old days, only raw data was in the input buffer, and any
552 * unused raw data in the buf would prevent the reading of socket data. */
553 static char *perform_io(size_t needed
, int flags
)
555 fd_set r_fds
, e_fds
, w_fds
;
558 size_t empty_buf_len
= 0;
562 if (iobuf
.in
.len
== 0 && iobuf
.in
.pos
!= 0) {
563 if (iobuf
.raw_input_ends_before
)
564 iobuf
.raw_input_ends_before
-= iobuf
.in
.pos
;
568 switch (flags
& PIO_NEED_FLAGS
) {
570 /* We never resize the circular input buffer. */
571 if (iobuf
.in
.size
< needed
) {
572 rprintf(FERROR
, "need to read %ld bytes, iobuf.in.buf is only %ld bytes.\n",
573 (long)needed
, (long)iobuf
.in
.size
);
574 exit_cleanup(RERR_PROTOCOL
);
577 if (DEBUG_GTE(IO
, 3)) {
578 rprintf(FINFO
, "[%s] perform_io(%ld, %sinput)\n",
579 who_am_i(), (long)needed
, flags
& PIO_CONSUME_INPUT
? "consume&" : "");
583 case PIO_NEED_OUTROOM
:
584 /* We never resize the circular output buffer. */
585 if (iobuf
.out
.size
- iobuf
.out_empty_len
< needed
) {
586 fprintf(stderr
, "need to write %ld bytes, iobuf.out.buf is only %ld bytes.\n",
587 (long)needed
, (long)(iobuf
.out
.size
- iobuf
.out_empty_len
));
588 exit_cleanup(RERR_PROTOCOL
);
591 if (DEBUG_GTE(IO
, 3)) {
592 rprintf(FINFO
, "[%s] perform_io(%ld, outroom) needs to flush %ld\n",
593 who_am_i(), (long)needed
,
594 iobuf
.out
.len
+ needed
> iobuf
.out
.size
595 ? (long)(iobuf
.out
.len
+ needed
- iobuf
.out
.size
) : 0L);
599 case PIO_NEED_MSGROOM
:
600 /* We never resize the circular message buffer. */
601 if (iobuf
.msg
.size
< needed
) {
602 fprintf(stderr
, "need to write %ld bytes, iobuf.msg.buf is only %ld bytes.\n",
603 (long)needed
, (long)iobuf
.msg
.size
);
604 exit_cleanup(RERR_PROTOCOL
);
607 if (DEBUG_GTE(IO
, 3)) {
608 rprintf(FINFO
, "[%s] perform_io(%ld, msgroom) needs to flush %ld\n",
609 who_am_i(), (long)needed
,
610 iobuf
.msg
.len
+ needed
> iobuf
.msg
.size
611 ? (long)(iobuf
.msg
.len
+ needed
- iobuf
.msg
.size
) : 0L);
616 if (DEBUG_GTE(IO
, 3))
617 rprintf(FINFO
, "[%s] perform_io(%ld, %d)\n", who_am_i(), (long)needed
, flags
);
621 exit_cleanup(RERR_UNSUPPORTED
);
625 switch (flags
& PIO_NEED_FLAGS
) {
627 if (iobuf
.in
.len
>= needed
)
630 case PIO_NEED_OUTROOM
:
631 /* Note that iobuf.out_empty_len doesn't factor into this check
632 * because iobuf.out.len already holds any needed header len. */
633 if (iobuf
.out
.len
+ needed
<= iobuf
.out
.size
)
636 case PIO_NEED_MSGROOM
:
637 if (iobuf
.msg
.len
+ needed
<= iobuf
.msg
.size
)
646 if (iobuf
.in_fd
>= 0 && iobuf
.in
.size
- iobuf
.in
.len
) {
647 if (!read_batch
|| batch_fd
>= 0) {
648 FD_SET(iobuf
.in_fd
, &r_fds
);
649 FD_SET(iobuf
.in_fd
, &e_fds
);
651 if (iobuf
.in_fd
> max_fd
)
652 max_fd
= iobuf
.in_fd
;
655 /* Only do more filesfrom processing if there is enough room in the out buffer. */
656 if (ff_forward_fd
>= 0 && iobuf
.out
.size
- iobuf
.out
.len
> FILESFROM_BUFLEN
*2) {
657 FD_SET(ff_forward_fd
, &r_fds
);
658 if (ff_forward_fd
> max_fd
)
659 max_fd
= ff_forward_fd
;
663 if (iobuf
.out_fd
>= 0) {
664 if (iobuf
.raw_flushing_ends_before
665 || (!iobuf
.msg
.len
&& iobuf
.out
.len
> iobuf
.out_empty_len
&& !(flags
& PIO_NEED_MSGROOM
))) {
666 if (OUT_MULTIPLEXED
&& !iobuf
.raw_flushing_ends_before
) {
667 /* The iobuf.raw_flushing_ends_before value can point off the end
668 * of the iobuf.out buffer for a while, for easier subtracting. */
669 iobuf
.raw_flushing_ends_before
= iobuf
.out
.pos
+ iobuf
.out
.len
;
671 SIVAL(iobuf
.out
.buf
+ iobuf
.raw_data_header_pos
, 0,
672 ((MPLEX_BASE
+ (int)MSG_DATA
)<<24) + iobuf
.out
.len
- 4);
674 if (DEBUG_GTE(IO
, 1)) {
675 rprintf(FINFO
, "[%s] send_msg(%d, %ld)\n",
676 who_am_i(), (int)MSG_DATA
, (long)iobuf
.out
.len
- 4);
679 /* reserve room for the next MSG_DATA header */
680 iobuf
.raw_data_header_pos
= iobuf
.raw_flushing_ends_before
;
681 if (iobuf
.raw_data_header_pos
>= iobuf
.out
.size
)
682 iobuf
.raw_data_header_pos
-= iobuf
.out
.size
;
683 else if (iobuf
.raw_data_header_pos
+ 4 > iobuf
.out
.size
) {
684 /* The 4-byte header won't fit at the end of the buffer,
685 * so we'll temporarily reduce the output buffer's size
686 * and put the header at the start of the buffer. */
687 reduce_iobuf_size(&iobuf
.out
, iobuf
.raw_data_header_pos
);
688 iobuf
.raw_data_header_pos
= 0;
690 /* Yes, it is possible for this to make len > size for a while. */
694 empty_buf_len
= iobuf
.out_empty_len
;
696 } else if (iobuf
.msg
.len
) {
702 FD_SET(iobuf
.out_fd
, &w_fds
);
703 if (iobuf
.out_fd
> max_fd
)
704 max_fd
= iobuf
.out_fd
;
710 switch (flags
& PIO_NEED_FLAGS
) {
713 if (kluge_around_eof
== 2)
715 if (iobuf
.in_fd
== -2)
716 whine_about_eof(True
);
717 rprintf(FERROR
, "error in perform_io: no fd for input.\n");
718 exit_cleanup(RERR_PROTOCOL
);
719 case PIO_NEED_OUTROOM
:
720 case PIO_NEED_MSGROOM
:
722 drain_multiplex_messages();
723 if (iobuf
.out_fd
== -2)
724 whine_about_eof(True
);
725 rprintf(FERROR
, "error in perform_io: no fd for output.\n");
726 exit_cleanup(RERR_PROTOCOL
);
728 /* No stated needs, so I guess this is OK. */
734 if (extra_flist_sending_enabled
) {
735 if (file_total
- file_old_total
< MAX_FILECNT_LOOKAHEAD
)
738 extra_flist_sending_enabled
= False
;
739 tv
.tv_sec
= select_timeout
;
742 tv
.tv_sec
= select_timeout
;
745 cnt
= select(max_fd
+ 1, &r_fds
, &w_fds
, &e_fds
, &tv
);
748 if (cnt
< 0 && errno
== EBADF
) {
750 exit_cleanup(RERR_SOCKETIO
);
752 if (extra_flist_sending_enabled
) {
753 extra_flist_sending_enabled
= False
;
754 send_extra_file_list(sock_f_out
, -1);
755 extra_flist_sending_enabled
= !flist_eof
;
757 check_timeout((flags
& PIO_NEED_INPUT
) != 0);
758 FD_ZERO(&r_fds
); /* Just in case... */
762 if (iobuf
.in_fd
>= 0 && FD_ISSET(iobuf
.in_fd
, &r_fds
)) {
763 size_t len
, pos
= iobuf
.in
.pos
+ iobuf
.in
.len
;
765 if (pos
>= iobuf
.in
.size
) {
766 pos
-= iobuf
.in
.size
;
767 len
= iobuf
.in
.size
- iobuf
.in
.len
;
769 len
= iobuf
.in
.size
- pos
;
770 if ((n
= read(iobuf
.in_fd
, iobuf
.in
.buf
+ pos
, len
)) <= 0) {
772 /* Signal that input has become invalid. */
773 if (!read_batch
|| batch_fd
< 0 || am_generator
)
778 if (errno
== EINTR
|| errno
== EWOULDBLOCK
|| errno
== EAGAIN
)
781 /* Don't write errors on a dead socket. */
782 if (iobuf
.in_fd
== sock_f_in
) {
785 rsyserr(FERROR_SOCKET
, errno
, "read error");
787 rsyserr(FERROR
, errno
, "read error");
788 exit_cleanup(RERR_SOCKETIO
);
791 if (msgs2stderr
&& DEBUG_GTE(IO
, 2))
792 rprintf(FINFO
, "[%s] recv=%ld\n", who_am_i(), (long)n
);
795 last_io_in
= time(NULL
);
796 if (flags
& PIO_NEED_INPUT
)
797 maybe_send_keepalive(last_io_in
, 0);
799 stats
.total_read
+= n
;
804 if (out
&& FD_ISSET(iobuf
.out_fd
, &w_fds
)) {
805 size_t len
= iobuf
.raw_flushing_ends_before
? iobuf
.raw_flushing_ends_before
- out
->pos
: out
->len
;
808 if (bwlimit_writemax
&& len
> bwlimit_writemax
)
809 len
= bwlimit_writemax
;
811 if (out
->pos
+ len
> out
->size
)
812 len
= out
->size
- out
->pos
;
813 if ((n
= write(iobuf
.out_fd
, out
->buf
+ out
->pos
, len
)) <= 0) {
814 if (errno
== EINTR
|| errno
== EWOULDBLOCK
|| errno
== EAGAIN
)
817 /* Don't write errors on a dead socket. */
820 iobuf
.out
.len
= iobuf
.msg
.len
= iobuf
.raw_flushing_ends_before
= 0;
821 rsyserr(FERROR_SOCKET
, errno
, "[%s] write error", who_am_i());
822 drain_multiplex_messages();
823 exit_cleanup(RERR_SOCKETIO
);
826 if (msgs2stderr
&& DEBUG_GTE(IO
, 2)) {
827 rprintf(FINFO
, "[%s] %s sent=%ld\n",
828 who_am_i(), out
== &iobuf
.out
? "out" : "msg", (long)n
);
832 last_io_out
= time(NULL
);
833 stats
.total_written
+= n
;
835 if (bwlimit_writemax
)
836 sleep_for_bwlimit(n
);
838 if ((out
->pos
+= n
) == out
->size
) {
839 if (iobuf
.raw_flushing_ends_before
)
840 iobuf
.raw_flushing_ends_before
-= out
->size
;
842 restore_iobuf_size(out
);
843 } else if (out
->pos
== iobuf
.raw_flushing_ends_before
)
844 iobuf
.raw_flushing_ends_before
= 0;
845 if ((out
->len
-= n
) == empty_buf_len
) {
847 restore_iobuf_size(out
);
849 iobuf
.raw_data_header_pos
= 0;
853 /* We need to help prevent deadlock by doing what reading
854 * we can whenever we are here trying to write. */
855 if (IN_MULTIPLEXED_AND_READY
&& !(flags
& PIO_NEED_INPUT
)) {
856 while (!iobuf
.raw_input_ends_before
&& iobuf
.in
.len
> 512)
858 if (flist_receiving_enabled
&& iobuf
.in
.len
> 512)
859 wait_for_receiver(); /* generator only */
862 if (ff_forward_fd
>= 0 && FD_ISSET(ff_forward_fd
, &r_fds
)) {
863 /* This can potentially flush all output and enable
864 * multiplexed output, so keep this last in the loop
865 * and be sure to not cache anything that would break
867 forward_filesfrom_data();
872 data
= iobuf
.in
.buf
+ iobuf
.in
.pos
;
874 if (flags
& PIO_CONSUME_INPUT
) {
875 iobuf
.in
.len
-= needed
;
876 iobuf
.in
.pos
+= needed
;
877 if (iobuf
.in
.pos
== iobuf
.raw_input_ends_before
)
878 iobuf
.raw_input_ends_before
= 0;
879 if (iobuf
.in
.pos
>= iobuf
.in
.size
) {
880 iobuf
.in
.pos
-= iobuf
.in
.size
;
881 if (iobuf
.raw_input_ends_before
)
882 iobuf
.raw_input_ends_before
-= iobuf
.in
.size
;
889 static void raw_read_buf(char *buf
, size_t len
)
891 size_t pos
= iobuf
.in
.pos
;
892 char *data
= perform_io(len
, PIO_INPUT_AND_CONSUME
);
893 if (iobuf
.in
.pos
<= pos
&& len
) {
894 size_t siz
= len
- iobuf
.in
.pos
;
895 memcpy(buf
, data
, siz
);
896 memcpy(buf
+ siz
, iobuf
.in
.buf
, iobuf
.in
.pos
);
898 memcpy(buf
, data
, len
);
901 static int32
raw_read_int(void)
904 if (iobuf
.in
.size
- iobuf
.in
.pos
>= 4)
905 data
= perform_io(4, PIO_INPUT_AND_CONSUME
);
907 raw_read_buf(data
= buf
, 4);
908 return IVAL(data
, 0);
911 void noop_io_until_death(void)
915 kluge_around_eof
= 2;
916 /* Setting an I/O timeout ensures that if something inexplicably weird
917 * happens, we won't hang around forever. */
922 read_buf(iobuf
.in_fd
, buf
, sizeof buf
);
925 /* Buffer a message for the multiplexed output stream. Is not used for (normal) MSG_DATA. */
926 int send_msg(enum msgcode code
, const char *buf
, size_t len
, int convert
)
930 BOOL want_debug
= DEBUG_GTE(IO
, 1) && convert
>= 0 && (msgs2stderr
|| code
!= MSG_INFO
);
932 if (!OUT_MULTIPLEXED
)
936 rprintf(FINFO
, "[%s] send_msg(%d, %ld)\n", who_am_i(), (int)code
, (long)len
);
938 /* When checking for enough free space for this message, we need to
939 * make sure that there is space for the 4-byte header, plus we'll
940 * assume that we may waste up to 3 bytes (if the header doesn't fit
941 * at the physical end of the buffer). */
943 if (convert
> 0 && ic_send
== (iconv_t
)-1)
946 /* Ensuring double-size room leaves space for maximal conversion expansion. */
947 needed
= len
*2 + 4 + 3;
950 needed
= len
+ 4 + 3;
951 if (iobuf
.msg
.len
+ needed
> iobuf
.msg
.size
)
952 perform_io(needed
, PIO_NEED_MSGROOM
);
954 pos
= iobuf
.msg
.pos
+ iobuf
.msg
.len
; /* Must be set after any flushing. */
955 if (pos
>= iobuf
.msg
.size
)
956 pos
-= iobuf
.msg
.size
;
957 else if (pos
+ 4 > iobuf
.msg
.size
) {
958 /* The 4-byte header won't fit at the end of the buffer,
959 * so we'll temporarily reduce the message buffer's size
960 * and put the header at the start of the buffer. */
961 reduce_iobuf_size(&iobuf
.msg
, pos
);
964 hdr
= iobuf
.msg
.buf
+ pos
;
966 iobuf
.msg
.len
+= 4; /* Allocate room for the coming header bytes. */
972 INIT_XBUF(inbuf
, (char*)buf
, len
, (size_t)-1);
975 iconvbufs(ic_send
, &inbuf
, &iobuf
.msg
,
976 ICB_INCLUDE_BAD
| ICB_INCLUDE_INCOMPLETE
| ICB_CIRCULAR_OUT
| ICB_INIT
);
978 rprintf(FERROR
, "overflowed iobuf.msg buffer in send_msg");
979 exit_cleanup(RERR_UNSUPPORTED
);
981 len
= iobuf
.msg
.len
- len
;
987 if ((pos
+= 4) == iobuf
.msg
.size
)
990 /* Handle a split copy if we wrap around the end of the circular buffer. */
991 if (pos
>= iobuf
.msg
.pos
&& (siz
= iobuf
.msg
.size
- pos
) < len
) {
992 memcpy(iobuf
.msg
.buf
+ pos
, buf
, siz
);
993 memcpy(iobuf
.msg
.buf
, buf
+ siz
, len
- siz
);
995 memcpy(iobuf
.msg
.buf
+ pos
, buf
, len
);
997 iobuf
.msg
.len
+= len
;
1000 SIVAL(hdr
, 0, ((MPLEX_BASE
+ (int)code
)<<24) + len
);
1002 if (want_debug
&& convert
> 0)
1003 rprintf(FINFO
, "[%s] converted msg len=%ld\n", who_am_i(), (long)len
);
1008 void send_msg_int(enum msgcode code
, int num
)
1012 if (DEBUG_GTE(IO
, 1))
1013 rprintf(FINFO
, "[%s] send_msg_int(%d, %d)\n", who_am_i(), (int)code
, num
);
1015 SIVAL(numbuf
, 0, num
);
1016 send_msg(code
, numbuf
, 4, -1);
1019 static void got_flist_entry_status(enum festatus status
, int ndx
)
1021 struct file_list
*flist
= flist_for_ndx(ndx
, "got_flist_entry_status");
1023 if (remove_source_files
) {
1025 active_bytecnt
-= F_LENGTH(flist
->files
[ndx
- flist
->ndx_start
]);
1029 flist
->in_progress
--;
1033 if (remove_source_files
)
1034 send_msg_int(MSG_SUCCESS
, ndx
);
1037 #ifdef SUPPORT_HARD_LINKS
1038 if (preserve_hard_links
) {
1039 struct file_struct
*file
= flist
->files
[ndx
- flist
->ndx_start
];
1040 if (F_IS_HLINKED(file
)) {
1041 if (status
== FES_NO_SEND
)
1042 flist_ndx_push(&hlink_list
, -2); /* indicates a failure follows */
1043 flist_ndx_push(&hlink_list
, ndx
);
1044 flist
->in_progress
++;
1052 flist
->in_progress
++;
1057 flist_ndx_push(&redo_list
, ndx
);
1062 /* Note the fds used for the main socket (which might really be a pipe
1063 * for a local transfer, but we can ignore that). */
1064 void io_set_sock_fds(int f_in
, int f_out
)
1070 void set_io_timeout(int secs
)
1073 allowed_lull
= (io_timeout
+ 1) / 2;
1075 if (!io_timeout
|| allowed_lull
> SELECT_TIMEOUT
)
1076 select_timeout
= SELECT_TIMEOUT
;
1078 select_timeout
= allowed_lull
;
1084 static void check_for_d_option_error(const char *msg
)
1086 static char rsync263_opts
[] = "BCDHIKLPRSTWabceghlnopqrtuvxz";
1091 || strncmp(msg
, REMOTE_OPTION_ERROR
, sizeof REMOTE_OPTION_ERROR
- 1) != 0)
1094 msg
+= sizeof REMOTE_OPTION_ERROR
- 1;
1095 if (*msg
== '-' || (colon
= strchr(msg
, ':')) == NULL
1096 || strncmp(colon
, REMOTE_OPTION_ERROR2
, sizeof REMOTE_OPTION_ERROR2
- 1) != 0)
1099 for ( ; *msg
!= ':'; msg
++) {
1102 else if (*msg
== 'e')
1104 else if (strchr(rsync263_opts
, *msg
) == NULL
)
1110 "*** Try using \"--old-d\" if remote rsync is <= 2.6.3 ***\n");
1114 /* This is used by the generator to limit how many file transfers can
1115 * be active at once when --remove-source-files is specified. Without
1116 * this, sender-side deletions were mostly happening at the end. */
1117 void increment_active_files(int ndx
, int itemizing
, enum logcode code
)
1120 /* TODO: tune these limits? */
1121 int limit
= active_bytecnt
>= 128*1024 ? 10 : 50;
1122 if (active_filecnt
< limit
)
1124 check_for_finished_files(itemizing
, code
, 0);
1125 if (active_filecnt
< limit
)
1127 wait_for_receiver();
1131 active_bytecnt
+= F_LENGTH(cur_flist
->files
[ndx
- cur_flist
->ndx_start
]);
1134 int get_redo_num(void)
1136 return flist_ndx_pop(&redo_list
);
1139 int get_hlink_num(void)
1141 return flist_ndx_pop(&hlink_list
);
1144 /* When we're the receiver and we have a local --files-from list of names
1145 * that needs to be sent over the socket to the sender, we have to do two
1146 * things at the same time: send the sender a list of what files we're
1147 * processing and read the incoming file+info list from the sender. We do
1148 * this by making recv_file_list() call forward_filesfrom_data(), which
1149 * will ensure that we forward data to the sender until we get some data
1150 * for recv_file_list() to use. */
1151 void start_filesfrom_forwarding(int fd
)
1153 if (protocol_version
< 31 && OUT_MULTIPLEXED
) {
1154 /* Older protocols send the files-from data w/o packaging
1155 * it in multiplexed I/O packets, so temporarily switch
1156 * to buffered I/O to match this behavior. */
1157 iobuf
.msg
.pos
= iobuf
.msg
.len
= 0; /* Be extra sure no messages go out. */
1158 ff_reenable_multiplex
= io_end_multiplex_out(MPLX_TO_BUFFERED
);
1162 alloc_xbuf(&ff_xb
, FILESFROM_BUFLEN
);
1165 /* Read a line into the "buf" buffer. */
1166 int read_line(int fd
, char *buf
, size_t bufsiz
, int flags
)
1171 if (flags
& RL_CONVERT
&& iconv_buf
.size
< bufsiz
)
1172 realloc_xbuf(&iconv_buf
, bufsiz
+ 1024);
1177 s
= flags
& RL_CONVERT
? iconv_buf
.buf
: buf
;
1181 eob
= s
+ bufsiz
- 1;
1183 /* We avoid read_byte() for files because files can return an EOF. */
1184 if (fd
== iobuf
.in_fd
)
1186 else if (safe_read(fd
, &ch
, 1) == 0)
1188 if (flags
& RL_EOL_NULLS
? ch
== '\0' : (ch
== '\r' || ch
== '\n')) {
1189 /* Skip empty lines if dumping comments. */
1190 if (flags
& RL_DUMP_COMMENTS
&& s
== buf
)
1199 if (flags
& RL_DUMP_COMMENTS
&& (*buf
== '#' || *buf
== ';'))
1203 if (flags
& RL_CONVERT
) {
1205 INIT_XBUF(outbuf
, buf
, 0, bufsiz
);
1207 iconv_buf
.len
= s
- iconv_buf
.buf
;
1208 iconvbufs(ic_recv
, &iconv_buf
, &outbuf
,
1209 ICB_INCLUDE_BAD
| ICB_INCLUDE_INCOMPLETE
| ICB_INIT
);
1210 outbuf
.buf
[outbuf
.len
] = '\0';
1218 void read_args(int f_in
, char *mod_name
, char *buf
, size_t bufsiz
, int rl_nulls
,
1219 char ***argv_p
, int *argc_p
, char **request_p
)
1221 int maxargs
= MAX_ARGS
;
1225 int rl_flags
= (rl_nulls
? RL_EOL_NULLS
: 0);
1228 rl_flags
|= (protect_args
&& ic_recv
!= (iconv_t
)-1 ? RL_CONVERT
: 0);
1231 if (!(argv
= new_array(char *, maxargs
)))
1232 out_of_memory("read_args");
1233 if (mod_name
&& !protect_args
)
1234 argv
[argc
++] = "rsyncd";
1237 if (read_line(f_in
, buf
, bufsiz
, rl_flags
) == 0)
1240 if (argc
== maxargs
-1) {
1241 maxargs
+= MAX_ARGS
;
1242 if (!(argv
= realloc_array(argv
, char *, maxargs
)))
1243 out_of_memory("read_args");
1248 *request_p
= strdup(buf
);
1252 glob_expand_module(mod_name
, buf
, &argv
, &argc
, &maxargs
);
1254 glob_expand(buf
, &argv
, &argc
, &maxargs
);
1256 if (!(p
= strdup(buf
)))
1257 out_of_memory("read_args");
1259 if (*p
== '.' && p
[1] == '\0')
1265 glob_expand(NULL
, NULL
, NULL
, NULL
);
1271 BOOL
io_start_buffering_out(int f_out
)
1273 if (msgs2stderr
&& DEBUG_GTE(IO
, 2))
1274 rprintf(FINFO
, "[%s] io_start_buffering_out(%d)\n", who_am_i(), f_out
);
1276 if (iobuf
.out
.buf
) {
1277 if (iobuf
.out_fd
== -1)
1278 iobuf
.out_fd
= f_out
;
1280 assert(f_out
== iobuf
.out_fd
);
1284 alloc_xbuf(&iobuf
.out
, ROUND_UP_1024(IO_BUFFER_SIZE
* 2));
1285 iobuf
.out_fd
= f_out
;
1290 BOOL
io_start_buffering_in(int f_in
)
1292 if (msgs2stderr
&& DEBUG_GTE(IO
, 2))
1293 rprintf(FINFO
, "[%s] io_start_buffering_in(%d)\n", who_am_i(), f_in
);
1296 if (iobuf
.in_fd
== -1)
1299 assert(f_in
== iobuf
.in_fd
);
1303 alloc_xbuf(&iobuf
.in
, ROUND_UP_1024(IO_BUFFER_SIZE
));
1309 void io_end_buffering_in(BOOL free_buffers
)
1311 if (msgs2stderr
&& DEBUG_GTE(IO
, 2)) {
1312 rprintf(FINFO
, "[%s] io_end_buffering_in(IOBUF_%s_BUFS)\n",
1313 who_am_i(), free_buffers
? "FREE" : "KEEP");
1317 free_xbuf(&iobuf
.in
);
1319 iobuf
.in
.pos
= iobuf
.in
.len
= 0;
1324 void io_end_buffering_out(BOOL free_buffers
)
1326 if (msgs2stderr
&& DEBUG_GTE(IO
, 2)) {
1327 rprintf(FINFO
, "[%s] io_end_buffering_out(IOBUF_%s_BUFS)\n",
1328 who_am_i(), free_buffers
? "FREE" : "KEEP");
1331 io_flush(FULL_FLUSH
);
1334 free_xbuf(&iobuf
.out
);
1335 free_xbuf(&iobuf
.msg
);
1341 void maybe_flush_socket(int important
)
1343 if (flist_eof
&& iobuf
.out
.buf
&& iobuf
.out
.len
> iobuf
.out_empty_len
1344 && (important
|| time(NULL
) - last_io_out
>= 5))
1345 io_flush(NORMAL_FLUSH
);
1348 /* Older rsync versions used to send either a MSG_NOOP (protocol 30) or a
1349 * raw-data-based keep-alive (protocol 29), both of which implied forwarding of
1350 * the message through the sender. Since the new timeout method does not need
1351 * any forwarding, we just send an empty MSG_DATA message, which works with all
1352 * rsync versions. This avoids any message forwarding, and leaves the raw-data
1353 * stream alone (since we can never be quite sure if that stream is in the
1354 * right state for a keep-alive message). */
1355 void maybe_send_keepalive(time_t now
, int flags
)
1357 if (flags
& MSK_ACTIVE_RECEIVER
)
1358 last_io_in
= now
; /* Fudge things when we're working hard on the files. */
1360 if (now
- last_io_out
>= allowed_lull
) {
1361 /* The receiver is special: it only sends keep-alive messages if it is
1362 * actively receiving data. Otherwise, it lets the generator timeout. */
1363 if (am_receiver
&& now
- last_io_in
>= io_timeout
)
1366 if (!iobuf
.msg
.len
&& iobuf
.out
.len
== iobuf
.out_empty_len
)
1367 send_msg(MSG_DATA
, "", 0, 0);
1368 if (!(flags
& MSK_ALLOW_FLUSH
)) {
1369 /* Let the caller worry about writing out the data. */
1370 } else if (iobuf
.msg
.len
)
1371 perform_io(iobuf
.msg
.size
- iobuf
.msg
.len
+ 1, PIO_NEED_MSGROOM
);
1372 else if (iobuf
.out
.len
> iobuf
.out_empty_len
)
1373 io_flush(NORMAL_FLUSH
);
1377 void start_flist_forward(int ndx
)
1379 write_int(iobuf
.out_fd
, ndx
);
1380 forward_flist_data
= 1;
1383 void stop_flist_forward(void)
1385 forward_flist_data
= 0;
1388 /* Read a message from a multiplexed source. */
1389 static void read_a_msg(void)
1391 char data
[BIGPATHBUFLEN
];
1395 /* This ensures that perform_io() does not try to do any message reading
1396 * until we've read all of the data for this message. We should also
1397 * try to avoid calling things that will cause data to be written via
1398 * perform_io() prior to this being reset to 1. */
1399 iobuf
.in_multiplexed
= -1;
1401 tag
= raw_read_int();
1403 msg_bytes
= tag
& 0xFFFFFF;
1404 tag
= (tag
>> 24) - MPLEX_BASE
;
1406 if (DEBUG_GTE(IO
, 1) && msgs2stderr
)
1407 rprintf(FINFO
, "[%s] got msg=%d, len=%ld\n", who_am_i(), (int)tag
, (long)msg_bytes
);
1411 assert(iobuf
.raw_input_ends_before
== 0);
1412 /* Though this does not yet read the data, we do mark where in
1413 * the buffer the msg data will end once it is read. It is
1414 * possible that this points off the end of the buffer, in
1415 * which case the gradual reading of the input stream will
1416 * cause this value to wrap around and eventually become real. */
1418 iobuf
.raw_input_ends_before
= iobuf
.in
.pos
+ msg_bytes
;
1419 iobuf
.in_multiplexed
= 1;
1422 if (msg_bytes
!= sizeof stats
.total_read
|| !am_generator
)
1424 raw_read_buf((char*)&stats
.total_read
, sizeof stats
.total_read
);
1425 iobuf
.in_multiplexed
= 1;
1428 if (msg_bytes
!= 4 || !am_generator
)
1430 val
= raw_read_int();
1431 iobuf
.in_multiplexed
= 1;
1432 got_flist_entry_status(FES_REDO
, val
);
1437 val
= raw_read_int();
1438 iobuf
.in_multiplexed
= 1;
1441 send_msg_int(MSG_IO_ERROR
, val
);
1443 case MSG_IO_TIMEOUT
:
1444 if (msg_bytes
!= 4 || am_server
|| am_generator
)
1446 val
= raw_read_int();
1447 iobuf
.in_multiplexed
= 1;
1448 if (!io_timeout
|| io_timeout
> val
) {
1449 if (INFO_GTE(MISC
, 2))
1450 rprintf(FINFO
, "Setting --timeout=%d to match server\n", val
);
1451 set_io_timeout(val
);
1455 /* Support protocol-30 keep-alive method. */
1458 iobuf
.in_multiplexed
= 1;
1460 maybe_send_keepalive(time(NULL
), MSK_ALLOW_FLUSH
);
1463 if (msg_bytes
>= sizeof data
)
1466 raw_read_buf(data
, msg_bytes
);
1467 iobuf
.in_multiplexed
= 1;
1468 send_msg(MSG_DELETED
, data
, msg_bytes
, 1);
1472 if (ic_recv
!= (iconv_t
)-1) {
1476 int flags
= ICB_INCLUDE_BAD
| ICB_INIT
;
1478 INIT_CONST_XBUF(outbuf
, data
);
1479 INIT_XBUF(inbuf
, ibuf
, 0, (size_t)-1);
1482 size_t len
= msg_bytes
> sizeof ibuf
- inbuf
.len
? sizeof ibuf
- inbuf
.len
: msg_bytes
;
1483 raw_read_buf(ibuf
+ inbuf
.len
, len
);
1486 if (!(msg_bytes
-= len
) && !ibuf
[inbuf
.len
-1])
1487 inbuf
.len
--, add_null
= 1;
1488 if (iconvbufs(ic_send
, &inbuf
, &outbuf
, flags
) < 0) {
1491 /* Buffer ended with an incomplete char, so move the
1492 * bytes to the start of the buffer and continue. */
1493 memmove(ibuf
, ibuf
+ inbuf
.pos
, inbuf
.len
);
1498 if (outbuf
.len
== outbuf
.size
)
1500 outbuf
.buf
[outbuf
.len
++] = '\0';
1502 msg_bytes
= outbuf
.len
;
1505 raw_read_buf(data
, msg_bytes
);
1506 iobuf
.in_multiplexed
= 1;
1507 /* A directory name was sent with the trailing null */
1508 if (msg_bytes
> 0 && !data
[msg_bytes
-1])
1509 log_delete(data
, S_IFDIR
);
1511 data
[msg_bytes
] = '\0';
1512 log_delete(data
, S_IFREG
);
1516 if (msg_bytes
!= 4) {
1518 rprintf(FERROR
, "invalid multi-message %d:%lu [%s%s]\n",
1519 tag
, (unsigned long)msg_bytes
, who_am_i(),
1520 inc_recurse
? "/inc" : "");
1521 exit_cleanup(RERR_STREAMIO
);
1523 val
= raw_read_int();
1524 iobuf
.in_multiplexed
= 1;
1526 got_flist_entry_status(FES_SUCCESS
, val
);
1528 successful_send(val
);
1533 val
= raw_read_int();
1534 iobuf
.in_multiplexed
= 1;
1536 got_flist_entry_status(FES_NO_SEND
, val
);
1538 send_msg_int(MSG_NO_SEND
, val
);
1540 case MSG_ERROR_SOCKET
:
1541 case MSG_ERROR_UTF8
:
1546 if (tag
== MSG_ERROR_SOCKET
)
1551 case MSG_ERROR_XFER
:
1553 if (msg_bytes
>= sizeof data
) {
1556 "multiplexing overflow %d:%lu [%s%s]\n",
1557 tag
, (unsigned long)msg_bytes
, who_am_i(),
1558 inc_recurse
? "/inc" : "");
1559 exit_cleanup(RERR_STREAMIO
);
1561 raw_read_buf(data
, msg_bytes
);
1562 iobuf
.in_multiplexed
= 1;
1563 rwrite((enum logcode
)tag
, data
, msg_bytes
, !am_generator
);
1564 if (first_message
) {
1565 if (list_only
&& !am_sender
&& tag
== 1 && msg_bytes
< sizeof data
) {
1566 data
[msg_bytes
] = '\0';
1567 check_for_d_option_error(data
);
1572 case MSG_ERROR_EXIT
:
1574 val
= raw_read_int();
1575 else if (msg_bytes
== 0)
1579 iobuf
.in_multiplexed
= 1;
1580 if (DEBUG_GTE(EXIT
, 3))
1581 rprintf(FINFO
, "[%s] got MSG_ERROR_EXIT with %ld bytes\n", who_am_i(), (long)msg_bytes
);
1582 if (msg_bytes
== 0) {
1583 if (!am_sender
&& !am_generator
) {
1584 if (DEBUG_GTE(EXIT
, 3)) {
1585 rprintf(FINFO
, "[%s] sending MSG_ERROR_EXIT (len 0)\n",
1588 send_msg(MSG_ERROR_EXIT
, "", 0, 0);
1589 io_flush(FULL_FLUSH
);
1591 } else if (protocol_version
>= 31) {
1593 if (DEBUG_GTE(EXIT
, 3)) {
1594 rprintf(FINFO
, "[%s] sending MSG_ERROR_EXIT with exit_code %d\n",
1597 send_msg_int(MSG_ERROR_EXIT
, val
);
1599 if (DEBUG_GTE(EXIT
, 3)) {
1600 rprintf(FINFO
, "[%s] sending MSG_ERROR_EXIT (len 0)\n",
1603 send_msg(MSG_ERROR_EXIT
, "", 0, 0);
1606 /* Send a negative linenum so that we don't end up
1607 * with a duplicate exit message. */
1608 _exit_cleanup(val
, __FILE__
, 0 - __LINE__
);
1610 rprintf(FERROR
, "unexpected tag %d [%s%s]\n",
1611 tag
, who_am_i(), inc_recurse
? "/inc" : "");
1612 exit_cleanup(RERR_STREAMIO
);
1615 assert(iobuf
.in_multiplexed
> 0);
1618 static void drain_multiplex_messages(void)
1620 while (IN_MULTIPLEXED_AND_READY
&& iobuf
.in
.len
) {
1621 if (iobuf
.raw_input_ends_before
) {
1622 size_t raw_len
= iobuf
.raw_input_ends_before
- iobuf
.in
.pos
;
1623 iobuf
.raw_input_ends_before
= 0;
1624 if (raw_len
>= iobuf
.in
.len
) {
1628 iobuf
.in
.len
-= raw_len
;
1629 if ((iobuf
.in
.pos
+= raw_len
) >= iobuf
.in
.size
)
1630 iobuf
.in
.pos
-= iobuf
.in
.size
;
1636 void wait_for_receiver(void)
1638 if (!iobuf
.raw_input_ends_before
)
1641 if (iobuf
.raw_input_ends_before
) {
1642 int ndx
= read_int(iobuf
.in_fd
);
1647 if (DEBUG_GTE(FLIST
, 3))
1648 rprintf(FINFO
, "[%s] flist_eof=1\n", who_am_i());
1654 exit_cleanup(RERR_STREAMIO
);
1657 struct file_list
*flist
;
1658 flist_receiving_enabled
= False
;
1659 if (DEBUG_GTE(FLIST
, 2)) {
1660 rprintf(FINFO
, "[%s] receiving flist for dir %d\n",
1663 flist
= recv_file_list(iobuf
.in_fd
);
1664 flist
->parent_ndx
= ndx
;
1665 #ifdef SUPPORT_HARD_LINKS
1666 if (preserve_hard_links
)
1667 match_hard_links(flist
);
1669 flist_receiving_enabled
= True
;
1674 unsigned short read_shortint(int f
)
1678 return (UVAL(b
, 1) << 8) + UVAL(b
, 0);
1681 int32
read_int(int f
)
1688 #if SIZEOF_INT32 > 4
1689 if (num
& (int32
)0x80000000)
1690 num
|= ~(int32
)0xffffffff;
1695 int32
read_varint(int f
)
1706 extra
= int_byte_extra
[ch
/ 4];
1708 uchar bit
= ((uchar
)1<<(8-extra
));
1709 if (extra
>= (int)sizeof u
.b
) {
1710 rprintf(FERROR
, "Overflow in read_varint()\n");
1711 exit_cleanup(RERR_STREAMIO
);
1713 read_buf(f
, u
.b
, extra
);
1714 u
.b
[extra
] = ch
& (bit
-1);
1717 #if CAREFUL_ALIGNMENT
1720 #if SIZEOF_INT32 > 4
1721 if (u
.x
& (int32
)0x80000000)
1722 u
.x
|= ~(int32
)0xffffffff;
1727 int64
read_varlong(int f
, uchar min_bytes
)
1736 #if SIZEOF_INT64 < 8
1741 read_buf(f
, b2
, min_bytes
);
1742 memcpy(u
.b
, b2
+1, min_bytes
-1);
1743 extra
= int_byte_extra
[CVAL(b2
, 0) / 4];
1745 uchar bit
= ((uchar
)1<<(8-extra
));
1746 if (min_bytes
+ extra
> (int)sizeof u
.b
) {
1747 rprintf(FERROR
, "Overflow in read_varlong()\n");
1748 exit_cleanup(RERR_STREAMIO
);
1750 read_buf(f
, u
.b
+ min_bytes
- 1, extra
);
1751 u
.b
[min_bytes
+ extra
- 1] = CVAL(b2
, 0) & (bit
-1);
1752 #if SIZEOF_INT64 < 8
1753 if (min_bytes
+ extra
> 5 || u
.b
[4] || CVAL(u
.b
,3) & 0x80) {
1754 rprintf(FERROR
, "Integer overflow: attempted 64-bit offset\n");
1755 exit_cleanup(RERR_UNSUPPORTED
);
1759 u
.b
[min_bytes
+ extra
- 1] = CVAL(b2
, 0);
1760 #if SIZEOF_INT64 < 8
1762 #elif CAREFUL_ALIGNMENT
1763 u
.x
= IVAL(u
.b
,0) | (((int64
)IVAL(u
.b
,4))<<32);
1768 int64
read_longint(int f
)
1770 #if SIZEOF_INT64 >= 8
1773 int32 num
= read_int(f
);
1775 if (num
!= (int32
)0xffffffff)
1778 #if SIZEOF_INT64 < 8
1779 rprintf(FERROR
, "Integer overflow: attempted 64-bit offset\n");
1780 exit_cleanup(RERR_UNSUPPORTED
);
1783 return IVAL(b
,0) | (((int64
)IVAL(b
,4))<<32);
1787 void read_buf(int f
, char *buf
, size_t len
)
1789 if (f
!= iobuf
.in_fd
) {
1790 if (safe_read(f
, buf
, len
) != len
)
1791 whine_about_eof(False
); /* Doesn't return. */
1795 if (!IN_MULTIPLEXED
) {
1796 raw_read_buf(buf
, len
);
1797 total_data_read
+= len
;
1798 if (forward_flist_data
)
1799 write_buf(iobuf
.out_fd
, buf
, len
);
1801 if (f
== write_batch_monitor_in
)
1802 safe_write(batch_fd
, buf
, len
);
1809 while (!iobuf
.raw_input_ends_before
)
1812 siz
= MIN(len
, iobuf
.raw_input_ends_before
- iobuf
.in
.pos
);
1813 if (siz
>= iobuf
.in
.size
)
1814 siz
= iobuf
.in
.size
;
1815 raw_read_buf(buf
, siz
);
1816 total_data_read
+= siz
;
1818 if (forward_flist_data
)
1819 write_buf(iobuf
.out_fd
, buf
, siz
);
1821 if (f
== write_batch_monitor_in
)
1822 safe_write(batch_fd
, buf
, siz
);
1824 if ((len
-= siz
) == 0)
1830 void read_sbuf(int f
, char *buf
, size_t len
)
1832 read_buf(f
, buf
, len
);
1836 uchar
read_byte(int f
)
1839 read_buf(f
, (char*)&c
, 1);
1843 int read_vstring(int f
, char *buf
, int bufsize
)
1845 int len
= read_byte(f
);
1848 len
= (len
& ~0x80) * 0x100 + read_byte(f
);
1850 if (len
>= bufsize
) {
1851 rprintf(FERROR
, "over-long vstring received (%d > %d)\n",
1857 read_buf(f
, buf
, len
);
1862 /* Populate a sum_struct with values from the socket. This is
1863 * called by both the sender and the receiver. */
1864 void read_sum_head(int f
, struct sum_struct
*sum
)
1866 int32 max_blength
= protocol_version
< 30 ? OLD_MAX_BLOCK_SIZE
: MAX_BLOCK_SIZE
;
1867 sum
->count
= read_int(f
);
1868 if (sum
->count
< 0) {
1869 rprintf(FERROR
, "Invalid checksum count %ld [%s]\n",
1870 (long)sum
->count
, who_am_i());
1871 exit_cleanup(RERR_PROTOCOL
);
1873 sum
->blength
= read_int(f
);
1874 if (sum
->blength
< 0 || sum
->blength
> max_blength
) {
1875 rprintf(FERROR
, "Invalid block length %ld [%s]\n",
1876 (long)sum
->blength
, who_am_i());
1877 exit_cleanup(RERR_PROTOCOL
);
1879 sum
->s2length
= protocol_version
< 27 ? csum_length
: (int)read_int(f
);
1880 if (sum
->s2length
< 0 || sum
->s2length
> MAX_DIGEST_LEN
) {
1881 rprintf(FERROR
, "Invalid checksum length %d [%s]\n",
1882 sum
->s2length
, who_am_i());
1883 exit_cleanup(RERR_PROTOCOL
);
1885 sum
->remainder
= read_int(f
);
1886 if (sum
->remainder
< 0 || sum
->remainder
> sum
->blength
) {
1887 rprintf(FERROR
, "Invalid remainder length %ld [%s]\n",
1888 (long)sum
->remainder
, who_am_i());
1889 exit_cleanup(RERR_PROTOCOL
);
1893 /* Send the values from a sum_struct over the socket. Set sum to
1894 * NULL if there are no checksums to send. This is called by both
1895 * the generator and the sender. */
1896 void write_sum_head(int f
, struct sum_struct
*sum
)
1898 static struct sum_struct null_sum
;
1903 write_int(f
, sum
->count
);
1904 write_int(f
, sum
->blength
);
1905 if (protocol_version
>= 27)
1906 write_int(f
, sum
->s2length
);
1907 write_int(f
, sum
->remainder
);
1910 /* Sleep after writing to limit I/O bandwidth usage.
1912 * @todo Rather than sleeping after each write, it might be better to
1913 * use some kind of averaging. The current algorithm seems to always
1914 * use a bit less bandwidth than specified, because it doesn't make up
1915 * for slow periods. But arguably this is a feature. In addition, we
1916 * ought to take the time used to write the data into account.
1918 * During some phases of big transfers (file FOO is uptodate) this is
1919 * called with a small bytes_written every time. As the kernel has to
1920 * round small waits up to guarantee that we actually wait at least the
1921 * requested number of microseconds, this can become grossly inaccurate.
1922 * We therefore keep track of the bytes we've written over time and only
1923 * sleep when the accumulated delay is at least 1 tenth of a second. */
1924 static void sleep_for_bwlimit(int bytes_written
)
1926 static struct timeval prior_tv
;
1927 static long total_written
= 0;
1928 struct timeval tv
, start_tv
;
1929 long elapsed_usec
, sleep_usec
;
1931 #define ONE_SEC 1000000L /* # of microseconds in a second */
1933 total_written
+= bytes_written
;
1935 gettimeofday(&start_tv
, NULL
);
1936 if (prior_tv
.tv_sec
) {
1937 elapsed_usec
= (start_tv
.tv_sec
- prior_tv
.tv_sec
) * ONE_SEC
1938 + (start_tv
.tv_usec
- prior_tv
.tv_usec
);
1939 total_written
-= elapsed_usec
* bwlimit
/ (ONE_SEC
/1024);
1940 if (total_written
< 0)
1944 sleep_usec
= total_written
* (ONE_SEC
/1024) / bwlimit
;
1945 if (sleep_usec
< ONE_SEC
/ 10) {
1946 prior_tv
= start_tv
;
1950 tv
.tv_sec
= sleep_usec
/ ONE_SEC
;
1951 tv
.tv_usec
= sleep_usec
% ONE_SEC
;
1952 select(0, NULL
, NULL
, NULL
, &tv
);
1954 gettimeofday(&prior_tv
, NULL
);
1955 elapsed_usec
= (prior_tv
.tv_sec
- start_tv
.tv_sec
) * ONE_SEC
1956 + (prior_tv
.tv_usec
- start_tv
.tv_usec
);
1957 total_written
= (sleep_usec
- elapsed_usec
) * bwlimit
/ (ONE_SEC
/1024);
1960 void io_flush(int flush_it_all
)
1962 if (iobuf
.out
.len
> iobuf
.out_empty_len
) {
1963 if (flush_it_all
) /* FULL_FLUSH: flush everything in the output buffers */
1964 perform_io(iobuf
.out
.size
- iobuf
.out_empty_len
, PIO_NEED_OUTROOM
);
1965 else /* NORMAL_FLUSH: flush at least 1 byte */
1966 perform_io(iobuf
.out
.size
- iobuf
.out
.len
+ 1, PIO_NEED_OUTROOM
);
1969 perform_io(iobuf
.msg
.size
, PIO_NEED_MSGROOM
);
1972 void write_shortint(int f
, unsigned short x
)
1976 b
[1] = (char)(x
>> 8);
1980 void write_int(int f
, int32 x
)
1987 void write_varint(int f
, int32 x
)
1995 while (cnt
> 1 && b
[cnt
] == 0)
1997 bit
= ((uchar
)1<<(7-cnt
+1));
1998 if (CVAL(b
, cnt
) >= bit
) {
2002 *b
= b
[cnt
] | ~(bit
*2-1);
2006 write_buf(f
, b
, cnt
);
2009 void write_varlong(int f
, int64 x
, uchar min_bytes
)
2016 #if SIZEOF_INT64 >= 8
2017 SIVAL(b
, 5, x
>> 32);
2019 if (x
<= 0x7FFFFFFF && x
>= 0)
2020 memset(b
+ 5, 0, 4);
2022 rprintf(FERROR
, "Integer overflow: attempted 64-bit offset\n");
2023 exit_cleanup(RERR_UNSUPPORTED
);
2027 while (cnt
> min_bytes
&& b
[cnt
] == 0)
2029 bit
= ((uchar
)1<<(7-cnt
+min_bytes
));
2030 if (CVAL(b
, cnt
) >= bit
) {
2033 } else if (cnt
> min_bytes
)
2034 *b
= b
[cnt
] | ~(bit
*2-1);
2038 write_buf(f
, b
, cnt
);
2042 * Note: int64 may actually be a 32-bit type if ./configure couldn't find any
2043 * 64-bit types on this platform.
2045 void write_longint(int f
, int64 x
)
2047 char b
[12], * const s
= b
+4;
2050 if (x
<= 0x7FFFFFFF && x
>= 0) {
2055 #if SIZEOF_INT64 < 8
2056 rprintf(FERROR
, "Integer overflow: attempted 64-bit offset\n");
2057 exit_cleanup(RERR_UNSUPPORTED
);
2060 SIVAL(s
, 4, x
>> 32);
2061 write_buf(f
, b
, 12);
2065 void write_buf(int f
, const char *buf
, size_t len
)
2069 if (f
!= iobuf
.out_fd
) {
2070 safe_write(f
, buf
, len
);
2074 if (iobuf
.out
.len
+ len
> iobuf
.out
.size
)
2075 perform_io(len
, PIO_NEED_OUTROOM
);
2077 pos
= iobuf
.out
.pos
+ iobuf
.out
.len
; /* Must be set after any flushing. */
2078 if (pos
>= iobuf
.out
.size
)
2079 pos
-= iobuf
.out
.size
;
2081 /* Handle a split copy if we wrap around the end of the circular buffer. */
2082 if (pos
>= iobuf
.out
.pos
&& (siz
= iobuf
.out
.size
- pos
) < len
) {
2083 memcpy(iobuf
.out
.buf
+ pos
, buf
, siz
);
2084 memcpy(iobuf
.out
.buf
, buf
+ siz
, len
- siz
);
2086 memcpy(iobuf
.out
.buf
+ pos
, buf
, len
);
2088 iobuf
.out
.len
+= len
;
2089 total_data_written
+= len
;
2092 if (f
== write_batch_monitor_out
)
2093 safe_write(batch_fd
, buf
, len
);
2096 /* Write a string to the connection */
2097 void write_sbuf(int f
, const char *buf
)
2099 write_buf(f
, buf
, strlen(buf
));
2102 void write_byte(int f
, uchar c
)
2104 write_buf(f
, (char *)&c
, 1);
2107 void write_vstring(int f
, const char *str
, int len
)
2109 uchar lenbuf
[3], *lb
= lenbuf
;
2114 "attempting to send over-long vstring (%d > %d)\n",
2116 exit_cleanup(RERR_PROTOCOL
);
2118 *lb
++ = len
/ 0x100 + 0x80;
2122 write_buf(f
, (char*)lenbuf
, lb
- lenbuf
+ 1);
2124 write_buf(f
, str
, len
);
2127 /* Send a file-list index using a byte-reduction method. */
2128 void write_ndx(int f
, int32 ndx
)
2130 static int32 prev_positive
= -1, prev_negative
= 1;
2131 int32 diff
, cnt
= 0;
2134 if (protocol_version
< 30 || read_batch
) {
2139 /* Send NDX_DONE as a single-byte 0 with no side effects. Send
2140 * negative nums as a positive after sending a leading 0xFF. */
2142 diff
= ndx
- prev_positive
;
2143 prev_positive
= ndx
;
2144 } else if (ndx
== NDX_DONE
) {
2149 b
[cnt
++] = (char)0xFF;
2151 diff
= ndx
- prev_negative
;
2152 prev_negative
= ndx
;
2155 /* A diff of 1 - 253 is sent as a one-byte diff; a diff of 254 - 32767
2156 * or 0 is sent as a 0xFE + a two-byte diff; otherwise we send 0xFE
2157 * & all 4 bytes of the (non-negative) num with the high-bit set. */
2158 if (diff
< 0xFE && diff
> 0)
2159 b
[cnt
++] = (char)diff
;
2160 else if (diff
< 0 || diff
> 0x7FFF) {
2161 b
[cnt
++] = (char)0xFE;
2162 b
[cnt
++] = (char)((ndx
>> 24) | 0x80);
2163 b
[cnt
++] = (char)ndx
;
2164 b
[cnt
++] = (char)(ndx
>> 8);
2165 b
[cnt
++] = (char)(ndx
>> 16);
2167 b
[cnt
++] = (char)0xFE;
2168 b
[cnt
++] = (char)(diff
>> 8);
2169 b
[cnt
++] = (char)diff
;
2171 write_buf(f
, b
, cnt
);
2174 /* Receive a file-list index using a byte-reduction method. */
2175 int32
read_ndx(int f
)
2177 static int32 prev_positive
= -1, prev_negative
= 1;
2178 int32
*prev_ptr
, num
;
2181 if (protocol_version
< 30)
2185 if (CVAL(b
, 0) == 0xFF) {
2187 prev_ptr
= &prev_negative
;
2188 } else if (CVAL(b
, 0) == 0)
2191 prev_ptr
= &prev_positive
;
2192 if (CVAL(b
, 0) == 0xFE) {
2194 if (CVAL(b
, 0) & 0x80) {
2195 b
[3] = CVAL(b
, 0) & ~0x80;
2197 read_buf(f
, b
+1, 2);
2200 num
= (UVAL(b
,0)<<8) + UVAL(b
,1) + *prev_ptr
;
2202 num
= UVAL(b
, 0) + *prev_ptr
;
2204 if (prev_ptr
== &prev_negative
)
2209 /* Read a line of up to bufsiz-1 characters into buf. Strips
2210 * the (required) trailing newline and all carriage returns.
2211 * Returns 1 for success; 0 for I/O error or truncation. */
2212 int read_line_old(int fd
, char *buf
, size_t bufsiz
, int eof_ok
)
2214 assert(fd
!= iobuf
.in_fd
);
2215 bufsiz
--; /* leave room for the null */
2216 while (bufsiz
> 0) {
2217 if (safe_read(fd
, buf
, 1) == 0) {
2235 void io_printf(int fd
, const char *format
, ...)
2238 char buf
[BIGPATHBUFLEN
];
2241 va_start(ap
, format
);
2242 len
= vsnprintf(buf
, sizeof buf
, format
, ap
);
2246 exit_cleanup(RERR_PROTOCOL
);
2248 if (len
> (int)sizeof buf
) {
2249 rprintf(FERROR
, "io_printf() was too long for the buffer.\n");
2250 exit_cleanup(RERR_PROTOCOL
);
2253 write_sbuf(fd
, buf
);
2256 /* Setup for multiplexing a MSG_* stream with the data stream. */
2257 void io_start_multiplex_out(int fd
)
2259 io_flush(FULL_FLUSH
);
2261 if (msgs2stderr
&& DEBUG_GTE(IO
, 2))
2262 rprintf(FINFO
, "[%s] io_start_multiplex_out(%d)\n", who_am_i(), fd
);
2265 alloc_xbuf(&iobuf
.msg
, ROUND_UP_1024(IO_BUFFER_SIZE
));
2267 iobuf
.out_empty_len
= 4; /* See also OUT_MULTIPLEXED */
2268 io_start_buffering_out(fd
);
2270 iobuf
.raw_data_header_pos
= iobuf
.out
.pos
+ iobuf
.out
.len
;
2274 /* Setup for multiplexing a MSG_* stream with the data stream. */
2275 void io_start_multiplex_in(int fd
)
2277 if (msgs2stderr
&& DEBUG_GTE(IO
, 2))
2278 rprintf(FINFO
, "[%s] io_start_multiplex_in(%d)\n", who_am_i(), fd
);
2280 iobuf
.in_multiplexed
= 1; /* See also IN_MULTIPLEXED */
2281 io_start_buffering_in(fd
);
2284 int io_end_multiplex_in(int mode
)
2286 int ret
= iobuf
.in_multiplexed
? iobuf
.in_fd
: -1;
2288 if (msgs2stderr
&& DEBUG_GTE(IO
, 2))
2289 rprintf(FINFO
, "[%s] io_end_multiplex_in(mode=%d)\n", who_am_i(), mode
);
2291 iobuf
.in_multiplexed
= 0;
2292 if (mode
== MPLX_SWITCHING
)
2293 iobuf
.raw_input_ends_before
= 0;
2295 assert(iobuf
.raw_input_ends_before
== 0);
2296 if (mode
!= MPLX_TO_BUFFERED
)
2297 io_end_buffering_in(mode
);
2302 int io_end_multiplex_out(int mode
)
2304 int ret
= iobuf
.out_empty_len
? iobuf
.out_fd
: -1;
2306 if (msgs2stderr
&& DEBUG_GTE(IO
, 2))
2307 rprintf(FINFO
, "[%s] io_end_multiplex_out(mode=%d)\n", who_am_i(), mode
);
2309 if (mode
!= MPLX_TO_BUFFERED
)
2310 io_end_buffering_out(mode
);
2312 io_flush(FULL_FLUSH
);
2315 iobuf
.out_empty_len
= 0;
2320 void start_write_batch(int fd
)
2322 /* Some communication has already taken place, but we don't
2323 * enable batch writing until here so that we can write a
2324 * canonical record of the communication even though the
2325 * actual communication so far depends on whether a daemon
2327 write_int(batch_fd
, protocol_version
);
2328 if (protocol_version
>= 30)
2329 write_byte(batch_fd
, compat_flags
);
2330 write_int(batch_fd
, checksum_seed
);
2333 write_batch_monitor_out
= fd
;
2335 write_batch_monitor_in
= fd
;
2338 void stop_write_batch(void)
2340 write_batch_monitor_out
= -1;
2341 write_batch_monitor_in
= -1;