/* * call-seq: * RepeatedField.[]=(index, value) * * Sets the element at the given index. On out-of-bounds assignments, extends * the array and fills the hole (if any) with default values. */ VALUE RepeatedField_index_set(VALUE _self, VALUE _index, VALUE val) { RepeatedField* self = ruby_to_RepeatedField(_self); upb_fieldtype_t field_type = self->field_type; VALUE field_type_class = self->field_type_class; int element_size = native_slot_size(field_type); int index = index_position(_index, self); if (index < 0 || index >= (INT_MAX - 1)) { return Qnil; } if (index >= self->size) { RepeatedField_reserve(self, index + 1); upb_fieldtype_t field_type = self->field_type; int element_size = native_slot_size(field_type); for (int i = self->size; i <= index; i++) { void* elem = (void *)(((uint8_t *)self->elements) + i * element_size); native_slot_init(field_type, elem); } self->size = index + 1; } void* memory = (void *) (((uint8_t *)self->elements) + index * element_size); native_slot_set(field_type, field_type_class, memory, val); return Qnil; }
/* * call-seq: * RepeatedField.==(other) => boolean * * Compares this repeated field to another. Repeated fields are equal if their * element types are equal, their lengths are equal, and each element is equal. * Elements are compared as per normal Ruby semantics, by calling their :== * methods (or performing a more efficient comparison for primitive types). * * Repeated fields with dissimilar element types are never equal, even if value * comparison (for example, between integers and floats) would have otherwise * indicated that every element has equal value. */ VALUE RepeatedField_eq(VALUE _self, VALUE _other) { if (_self == _other) { return Qtrue; } RepeatedField* self = ruby_to_RepeatedField(_self); if (TYPE(_other) == T_ARRAY) { VALUE self_ary = RepeatedField_to_ary(_self); return rb_equal(self_ary, _other); } RepeatedField* other = ruby_to_RepeatedField(_other); if (self->field_type != other->field_type || self->field_type_class != other->field_type_class || self->size != other->size) { return Qfalse; } upb_fieldtype_t field_type = self->field_type; size_t elem_size = native_slot_size(field_type); size_t off = 0; for (int i = 0; i < self->size; i++, off += elem_size) { void* self_mem = ((uint8_t *)self->elements) + off; void* other_mem = ((uint8_t *)other->elements) + off; if (!native_slot_eq(field_type, self_mem, other_mem)) { return Qfalse; } } return Qtrue; }
static VALUE table_key(Map* self, VALUE key, char* buf, const char** out_key, size_t* out_length) { switch (self->key_type) { case UPB_TYPE_BYTES: case UPB_TYPE_STRING: // Strings: use string content directly. Check_Type(key, T_STRING); key = native_slot_encode_and_freeze_string(self->key_type, key); *out_key = RSTRING_PTR(key); *out_length = RSTRING_LEN(key); break; case UPB_TYPE_BOOL: case UPB_TYPE_INT32: case UPB_TYPE_INT64: case UPB_TYPE_UINT32: case UPB_TYPE_UINT64: native_slot_set(self->key_type, Qnil, buf, key); *out_key = buf; *out_length = native_slot_size(self->key_type); break; default: // Map constructor should not allow a Map with another key type to be // constructed. assert(false); break; } return key; }
void RepeatedField_mark(void* _self) { RepeatedField* self = (RepeatedField*)_self; rb_gc_mark(self->field_type_class); upb_fieldtype_t field_type = self->field_type; int element_size = native_slot_size(field_type); for (int i = 0; i < self->size; i++) { void* memory = (((uint8_t *)self->elements) + i * element_size); native_slot_mark(self->field_type, memory); } }
// Used by parsing handlers. void RepeatedField_push_native(VALUE _self, void* data) { RepeatedField* self = ruby_to_RepeatedField(_self); upb_fieldtype_t field_type = self->field_type; int element_size = native_slot_size(field_type); RepeatedField_reserve(self, self->size + 1); int index = self->size; void* memory = (void *) (((uint8_t *)self->elements) + index * element_size); memcpy(memory, data, element_size); self->size++; }
/* * call-seq: * RepeatedField.push(value) * * Adds a new element to the repeated field. */ VALUE RepeatedField_push(VALUE _self, VALUE val) { RepeatedField* self = ruby_to_RepeatedField(_self); upb_fieldtype_t field_type = self->field_type; int element_size = native_slot_size(field_type); RepeatedField_reserve(self, self->size + 1); int index = self->size; void* memory = (void *) (((uint8_t *)self->elements) + index * element_size); native_slot_set(field_type, self->field_type_class, memory, val); // native_slot_set may raise an error; bump index only after set. self->size++; return _self; }
/* * call-seq: * RepeatedField.[](index) => value * * Accesses the element at the given index. Returns nil on out-of-bounds */ VALUE RepeatedField_index(VALUE _self, VALUE _index) { RepeatedField* self = ruby_to_RepeatedField(_self); int element_size = native_slot_size(self->field_type); upb_fieldtype_t field_type = self->field_type; VALUE field_type_class = self->field_type_class; int index = index_position(_index, self); if (index < 0 || index >= self->size) { return Qnil; } void* memory = (void *) (((uint8_t *)self->elements) + index * element_size); return native_slot_get(field_type, field_type_class, memory); }
/* * call-seq: * RepeatedField.to_ary => array * * Used when converted implicitly into array, e.g. compared to an Array. * Also called as a fallback of Object#to_a */ VALUE RepeatedField_to_ary(VALUE _self) { RepeatedField* self = ruby_to_RepeatedField(_self); upb_fieldtype_t field_type = self->field_type; size_t elem_size = native_slot_size(field_type); size_t off = 0; VALUE ary = rb_ary_new2(self->size); for (int i = 0; i < self->size; i++, off += elem_size) { void* mem = ((uint8_t *)self->elements) + off; VALUE elem = native_slot_get(field_type, self->field_type_class, mem); rb_ary_push(ary, elem); } return ary; }
/* * call-seq: * RepeatedField.[](index) => value * * Accesses the element at the given index. Throws an exception on out-of-bounds * errors. */ VALUE RepeatedField_index(VALUE _self, VALUE _index) { RepeatedField* self = ruby_to_RepeatedField(_self); int element_size = native_slot_size(self->field_type); upb_fieldtype_t field_type = self->field_type; VALUE field_type_class = self->field_type_class; int index = NUM2INT(_index); if (index < 0 || index >= self->size) { rb_raise(rb_eRangeError, "Index out of range"); } void* memory = (void *) (((uint8_t *)self->elements) + index * element_size); return native_slot_get(field_type, field_type_class, memory); }
bool native_slot_eq(upb_fieldtype_t type, void* mem1, void* mem2) { switch (type) { case UPB_TYPE_STRING: case UPB_TYPE_BYTES: case UPB_TYPE_MESSAGE: { VALUE val1 = DEREF(mem1, VALUE); VALUE val2 = DEREF(mem2, VALUE); VALUE ret = rb_funcall(val1, rb_intern("=="), 1, val2); return ret == Qtrue; } default: return !memcmp(mem1, mem2, native_slot_size(type)); } }
/* * call-seq: * RepeatedField.pop => value * * Removes the last element and returns it. Throws an exception if the repeated * field is empty. */ VALUE RepeatedField_pop(VALUE _self) { RepeatedField* self = ruby_to_RepeatedField(_self); upb_fieldtype_t field_type = self->field_type; VALUE field_type_class = self->field_type_class; int element_size = native_slot_size(field_type); if (self->size == 0) { rb_raise(rb_eRangeError, "Pop from empty repeated field is not allowed."); } int index = self->size - 1; void* memory = (void *) (((uint8_t *)self->elements) + index * element_size); VALUE ret = native_slot_get(field_type, field_type_class, memory); self->size--; return ret; }
/* * call-seq: * RepeatedField.each(&block) * * Invokes the block once for each element of the repeated field. RepeatedField * also includes Enumerable; combined with this method, the repeated field thus * acts like an ordinary Ruby sequence. */ VALUE RepeatedField_each(VALUE _self) { RepeatedField* self = ruby_to_RepeatedField(_self); upb_fieldtype_t field_type = self->field_type; VALUE field_type_class = self->field_type_class; int element_size = native_slot_size(field_type); size_t off = 0; for (int i = 0; i < self->size; i++, off += element_size) { void* memory = (void *) (((uint8_t *)self->elements) + off); VALUE val = native_slot_get(field_type, field_type_class, memory); rb_yield(val); } return _self; }
/* * Private ruby method, used by RepeatedField.pop */ VALUE RepeatedField_pop_one(VALUE _self) { RepeatedField* self = ruby_to_RepeatedField(_self); upb_fieldtype_t field_type = self->field_type; VALUE field_type_class = self->field_type_class; int element_size = native_slot_size(field_type); if (self->size == 0) { return Qnil; } int index = self->size - 1; void* memory = (void *) (((uint8_t *)self->elements) + index * element_size); VALUE ret = native_slot_get(field_type, field_type_class, memory); self->size--; return ret; }
VALUE RepeatedField_subarray(VALUE _self, long beg, long len) { RepeatedField* self = ruby_to_RepeatedField(_self); int element_size = native_slot_size(self->field_type); upb_fieldtype_t field_type = self->field_type; VALUE field_type_class = self->field_type_class; size_t off = beg * element_size; VALUE ary = rb_ary_new2(len); for (int i = beg; i < beg + len; i++, off += element_size) { void* mem = ((uint8_t *)self->elements) + off; VALUE elem = native_slot_get(field_type, field_type_class, mem); rb_ary_push(ary, elem); } return ary; }
// Internal only: used by Google::Protobuf.deep_copy. VALUE RepeatedField_deep_copy(VALUE _self) { RepeatedField* self = ruby_to_RepeatedField(_self); VALUE new_rptfield = RepeatedField_new_this_type(_self); RepeatedField* new_rptfield_self = ruby_to_RepeatedField(new_rptfield); RepeatedField_reserve(new_rptfield_self, self->size); upb_fieldtype_t field_type = self->field_type; size_t elem_size = native_slot_size(field_type); size_t off = 0; for (int i = 0; i < self->size; i++, off += elem_size) { void* to_mem = (uint8_t *)new_rptfield_self->elements + off; void* from_mem = (uint8_t *)self->elements + off; native_slot_deep_copy(field_type, to_mem, from_mem); new_rptfield_self->size++; } return new_rptfield; }
void RepeatedField_reserve(RepeatedField* self, int new_size) { if (new_size <= self->capacity) { return; } if (self->capacity == 0) { self->capacity = kInitialSize; } while (self->capacity < new_size) { self->capacity *= 2; } void* old_elems = self->elements; int elem_size = native_slot_size(self->field_type); self->elements = ALLOC_N(uint8_t, elem_size * self->capacity); if (old_elems != NULL) { memcpy(self->elements, old_elems, self->size * elem_size); xfree(old_elems); } }
void native_slot_deep_copy(upb_fieldtype_t type, void* to, void* from) { switch (type) { case UPB_TYPE_STRING: case UPB_TYPE_BYTES: { VALUE from_val = DEREF(from, VALUE); DEREF(to, VALUE) = (from_val != Qnil) ? rb_funcall(from_val, rb_intern("dup"), 0) : Qnil; break; } case UPB_TYPE_MESSAGE: { VALUE from_val = DEREF(from, VALUE); DEREF(to, VALUE) = (from_val != Qnil) ? Message_deep_copy(from_val) : Qnil; break; } default: memcpy(to, from, native_slot_size(type)); } }
/* * call-seq: * RepeatedField.[](index) => value * * Accesses the element at the given index. Returns nil on out-of-bounds */ VALUE RepeatedField_index(int argc, VALUE* argv, VALUE _self) { RepeatedField* self = ruby_to_RepeatedField(_self); int element_size = native_slot_size(self->field_type); upb_fieldtype_t field_type = self->field_type; VALUE field_type_class = self->field_type_class; VALUE arg = argv[0]; long beg, len; if (argc == 1){ if (FIXNUM_P(arg)) { /* standard case */ int index = index_position(argv[0], self); if (index < 0 || index >= self->size) { return Qnil; } void* memory = (void *) (((uint8_t *)self->elements) + index * element_size); return native_slot_get(field_type, field_type_class, memory); }else{ /* check if idx is Range */ size_t off; switch (rb_range_beg_len(arg, &beg, &len, self->size, 0)) { case Qfalse: break; case Qnil: return Qnil; default: return RepeatedField_subarray(_self, beg, len); } } } /* assume 2 arguments */ beg = NUM2LONG(argv[0]); len = NUM2LONG(argv[1]); if (beg < 0) { beg += self->size; } if (beg >= self->size) { return Qnil; } return RepeatedField_subarray(_self, beg, len); }
/* * call-seq: * RepeatedField.hash => hash_value * * Returns a hash value computed from this repeated field's elements. */ VALUE RepeatedField_hash(VALUE _self) { RepeatedField* self = ruby_to_RepeatedField(_self); VALUE hash = LL2NUM(0); upb_fieldtype_t field_type = self->field_type; VALUE field_type_class = self->field_type_class; size_t elem_size = native_slot_size(field_type); size_t off = 0; for (int i = 0; i < self->size; i++, off += elem_size) { void* mem = ((uint8_t *)self->elements) + off; VALUE elem = native_slot_get(field_type, field_type_class, mem); hash = rb_funcall(hash, rb_intern("<<"), 1, INT2NUM(2)); hash = rb_funcall(hash, rb_intern("^"), 1, rb_funcall(elem, rb_intern("hash"), 0)); } return hash; }
MessageLayout* create_layout(const upb_msgdef* msgdef) { MessageLayout* layout = ALLOC(MessageLayout); int nfields = upb_msgdef_numfields(msgdef); upb_msg_field_iter it; upb_msg_oneof_iter oit; size_t off = 0; layout->fields = ALLOC_N(MessageField, nfields); size_t hasbit = 0; for (upb_msg_field_begin(&it, msgdef); !upb_msg_field_done(&it); upb_msg_field_next(&it)) { const upb_fielddef* field = upb_msg_iter_field(&it); if (upb_fielddef_haspresence(field)) { layout->fields[upb_fielddef_index(field)].hasbit = hasbit++; } else { layout->fields[upb_fielddef_index(field)].hasbit = MESSAGE_FIELD_NO_HASBIT; } } if (hasbit != 0) { off += (hasbit + 8 - 1) / 8; } for (upb_msg_field_begin(&it, msgdef); !upb_msg_field_done(&it); upb_msg_field_next(&it)) { const upb_fielddef* field = upb_msg_iter_field(&it); size_t field_size; if (upb_fielddef_containingoneof(field)) { // Oneofs are handled separately below. continue; } // Allocate |field_size| bytes for this field in the layout. field_size = 0; if (upb_fielddef_label(field) == UPB_LABEL_REPEATED) { field_size = sizeof(VALUE); } else { field_size = native_slot_size(upb_fielddef_type(field)); } // Align current offset up to |size| granularity. off = align_up_to(off, field_size); layout->fields[upb_fielddef_index(field)].offset = off; layout->fields[upb_fielddef_index(field)].case_offset = MESSAGE_FIELD_NO_CASE; off += field_size; } // Handle oneofs now -- we iterate over oneofs specifically and allocate only // one slot per oneof. // // We assign all value slots first, then pack the 'case' fields at the end, // since in the common case (modern 64-bit platform) these are 8 bytes and 4 // bytes respectively and we want to avoid alignment overhead. // // Note that we reserve 4 bytes (a uint32) per 'case' slot because the value // space for oneof cases is conceptually as wide as field tag numbers. In // practice, it's unlikely that a oneof would have more than e.g. 256 or 64K // members (8 or 16 bits respectively), so conceivably we could assign // consecutive case numbers and then pick a smaller oneof case slot size, but // the complexity to implement this indirection is probably not worthwhile. for (upb_msg_oneof_begin(&oit, msgdef); !upb_msg_oneof_done(&oit); upb_msg_oneof_next(&oit)) { const upb_oneofdef* oneof = upb_msg_iter_oneof(&oit); upb_oneof_iter fit; // Always allocate NATIVE_SLOT_MAX_SIZE bytes, but share the slot between // all fields. size_t field_size = NATIVE_SLOT_MAX_SIZE; // Align the offset. off = align_up_to(off, field_size); // Assign all fields in the oneof this same offset. for (upb_oneof_begin(&fit, oneof); !upb_oneof_done(&fit); upb_oneof_next(&fit)) { const upb_fielddef* field = upb_oneof_iter_field(&fit); layout->fields[upb_fielddef_index(field)].offset = off; } off += field_size; } // Now the case fields. for (upb_msg_oneof_begin(&oit, msgdef); !upb_msg_oneof_done(&oit); upb_msg_oneof_next(&oit)) { const upb_oneofdef* oneof = upb_msg_iter_oneof(&oit); upb_oneof_iter fit; size_t field_size = sizeof(uint32_t); // Align the offset. off = (off + field_size - 1) & ~(field_size - 1); // Assign all fields in the oneof this same offset. for (upb_oneof_begin(&fit, oneof); !upb_oneof_done(&fit); upb_oneof_next(&fit)) { const upb_fielddef* field = upb_oneof_iter_field(&fit); layout->fields[upb_fielddef_index(field)].case_offset = off; } off += field_size; } layout->size = off; layout->msgdef = msgdef; upb_msgdef_ref(layout->msgdef, &layout->msgdef); return layout; }
void native_slot_dup(upb_fieldtype_t type, void* to, void* from) { memcpy(to, from, native_slot_size(type)); }
void* RepeatedField_index_native(VALUE _self, int index) { RepeatedField* self = ruby_to_RepeatedField(_self); upb_fieldtype_t field_type = self->field_type; int element_size = native_slot_size(field_type); return ((uint8_t *)self->elements) + index * element_size; }