This repository has been archived on 2023-10-09. You can view files and clone it. You cannot open issues or pull requests or push a commit.
Files
blender-archive/source/blender/editors/mesh/editmesh_undo.c
Monique Dewanchand 68589a31eb ViewLayer: Lazy sync of scene data.
When a change happens which invalidates view layers the syncing will be postponed until the first usage.
This will improve importing or adding many objects in a single operation/script.

`BKE_view_layer_need_resync_tag` is used to tag the view layer to be out of sync. Before accessing
`BKE_view_layer_active_base_get`, `BKE_view_layer_active_object_get`, `BKE_view_layer_active_collection`
or `BKE_view_layer_object_bases` the caller should call `BKE_view_layer_synced_ensure`.

Having two functions ensures that partial syncing could be added as smaller patches in the future. Tagging a
view layer out of sync could be replaced with a partial sync. Eventually the number of full resyncs could be
reduced. After all tagging has been replaced with partial syncs the ensure_sync could be phased out.

This patch has been added to discuss the details and consequences of the current approach. For clarity
the call to BKE_view_layer_ensure_sync is placed close to the getters.
In the future this could be placed in more strategical places to reduce the number of calls or improve
performance. Finding those strategical places isn't that clear. When multiple operations are grouped
in a single script you might want to always check for resync.

Some areas found that can be improved. This list isn't complete.
These areas aren't addressed by this patch as these changes would be hard to detect to the reviewer.
The idea is to add changes to these areas as a separate patch. It might be that the initial commit would reduce
performance compared to master, but will be fixed by the additional patches.

**Object duplication**
During object duplication the syncing is temporarily disabled. With this patch this isn't useful as when disabled
the view_layer is accessed to locate bases. This can be improved by first locating the source bases, then duplicate
and sync and locate the new bases. Will be solved in a separate patch for clarity reasons ({D15886}).

**Object add**
`BKE_object_add` not only adds a new object, but also selects and activates the new base. This requires the
view_layer to be resynced. Some callers reverse the selection and activation (See `get_new_constraint_target`).
We should make the selection and activation optional. This would make it possible to add multiple objects
without having to resync per object.

**Postpone Activate Base**
Setting the basact is done in many locations. They follow a rule as after an action find the base and set
the basact. Finding the base could require a resync. The idea is to store in the view_layer the object which
base will be set in the basact during the next sync, reducing the times resyncing needs to happen.

Reviewed By: mont29

Maniphest Tasks: T73411

Differential Revision: https://developer.blender.org/D15885
2022-09-14 21:34:38 +02:00

906 lines
26 KiB
C

/* SPDX-License-Identifier: GPL-2.0-or-later */
/** \file
* \ingroup edmesh
*/
#include "MEM_guardedalloc.h"
#include "CLG_log.h"
#include "DNA_key_types.h"
#include "DNA_layer_types.h"
#include "DNA_mesh_types.h"
#include "DNA_meshdata_types.h"
#include "DNA_object_types.h"
#include "DNA_scene_types.h"
#include "BLI_array_utils.h"
#include "BLI_listbase.h"
#include "BKE_context.h"
#include "BKE_customdata.h"
#include "BKE_editmesh.h"
#include "BKE_key.h"
#include "BKE_layer.h"
#include "BKE_lib_id.h"
#include "BKE_main.h"
#include "BKE_mesh.h"
#include "BKE_object.h"
#include "BKE_undo_system.h"
#include "DEG_depsgraph.h"
#include "ED_mesh.h"
#include "ED_object.h"
#include "ED_undo.h"
#include "ED_util.h"
#include "WM_api.h"
#include "WM_types.h"
#define USE_ARRAY_STORE
#ifdef USE_ARRAY_STORE
// # define DEBUG_PRINT
// # define DEBUG_TIME
# ifdef DEBUG_TIME
# include "PIL_time_utildefines.h"
# endif
# include "BLI_array_store.h"
# include "BLI_array_store_utils.h"
/* check on best size later... */
# define ARRAY_CHUNK_SIZE 256
# define USE_ARRAY_STORE_THREAD
#endif
#ifdef USE_ARRAY_STORE_THREAD
# include "BLI_task.h"
#endif
/** We only need this locally. */
static CLG_LogRef LOG = {"ed.undo.mesh"};
/* -------------------------------------------------------------------- */
/** \name Undo Conversion
* \{ */
#ifdef USE_ARRAY_STORE
/* Single linked list of layers stored per type */
typedef struct BArrayCustomData {
struct BArrayCustomData *next;
eCustomDataType type;
int states_len; /* number of layers for each type */
BArrayState *states[0];
} BArrayCustomData;
#endif
typedef struct UndoMesh {
/**
* This undo-meshes in `um_arraystore.local_links`.
* Not to be confused with the next and previous undo steps.
*/
struct UndoMesh *local_next, *local_prev;
Mesh me;
int selectmode;
char uv_selectmode;
/** \note
* This isn't a perfect solution, if you edit keys and change shapes this works well
* (fixing T32442), but editing shape keys, going into object mode, removing or changing their
* order, then go back into editmode and undo will give issues - where the old index will be
* out of sync with the new object index.
*
* There are a few ways this could be made to work but for now its a known limitation with mixing
* object and editmode operations - Campbell. */
int shapenr;
#ifdef USE_ARRAY_STORE
/* NULL arrays are considered empty */
struct { /* most data is stored as 'custom' data */
BArrayCustomData *vdata, *edata, *ldata, *pdata;
BArrayState **keyblocks;
BArrayState *mselect;
} store;
#endif /* USE_ARRAY_STORE */
size_t undo_size;
} UndoMesh;
#ifdef USE_ARRAY_STORE
/* -------------------------------------------------------------------- */
/** \name Array Store
* \{ */
static struct {
struct BArrayStore_AtSize bs_stride;
int users;
/**
* A list of #UndoMesh items ordered from oldest to newest
* used to access previous undo data for a mesh.
*/
ListBase local_links;
# ifdef USE_ARRAY_STORE_THREAD
TaskPool *task_pool;
# endif
} um_arraystore = {{NULL}};
static void um_arraystore_cd_compact(struct CustomData *cdata,
const size_t data_len,
bool create,
const BArrayCustomData *bcd_reference,
BArrayCustomData **r_bcd_first)
{
if (data_len == 0) {
if (create) {
*r_bcd_first = NULL;
}
}
const BArrayCustomData *bcd_reference_current = bcd_reference;
BArrayCustomData *bcd = NULL, *bcd_first = NULL, *bcd_prev = NULL;
for (int layer_start = 0, layer_end; layer_start < cdata->totlayer; layer_start = layer_end) {
const eCustomDataType type = cdata->layers[layer_start].type;
/* Perform a full copy on dynamic layers.
*
* Unfortunately we can't compare dynamic layer types as they contain allocated pointers,
* which burns CPU cycles looking for duplicate data that doesn't exist.
* The array data isn't comparable once copied from the mesh,
* this bottlenecks on high poly meshes, see T84114.
*
* Notes:
*
* - Ideally the data would be expanded into a format that could be de-duplicated effectively,
* this would require a flat representation of each dynamic custom-data layer.
*
* - The data in the layer could be kept as-is to save on the extra copy,
* it would complicate logic in this function.
*/
const bool layer_type_is_dynamic = CustomData_layertype_is_dynamic(type);
layer_end = layer_start + 1;
while ((layer_end < cdata->totlayer) && (type == cdata->layers[layer_end].type)) {
layer_end++;
}
const int stride = CustomData_sizeof(type);
BArrayStore *bs = create ? BLI_array_store_at_size_ensure(
&um_arraystore.bs_stride, stride, ARRAY_CHUNK_SIZE) :
NULL;
const int layer_len = layer_end - layer_start;
if (create) {
if (bcd_reference_current && (bcd_reference_current->type == type)) {
/* common case, the reference is aligned */
}
else {
bcd_reference_current = NULL;
/* Do a full lookup when unaligned. */
if (bcd_reference) {
const BArrayCustomData *bcd_iter = bcd_reference;
while (bcd_iter) {
if (bcd_iter->type == type) {
bcd_reference_current = bcd_iter;
break;
}
bcd_iter = bcd_iter->next;
}
}
}
}
if (create) {
bcd = MEM_callocN(sizeof(BArrayCustomData) + (layer_len * sizeof(BArrayState *)), __func__);
bcd->next = NULL;
bcd->type = type;
bcd->states_len = layer_end - layer_start;
if (bcd_prev) {
bcd_prev->next = bcd;
bcd_prev = bcd;
}
else {
bcd_first = bcd;
bcd_prev = bcd;
}
}
CustomDataLayer *layer = &cdata->layers[layer_start];
for (int i = 0; i < layer_len; i++, layer++) {
if (create) {
if (layer->data) {
BArrayState *state_reference = (bcd_reference_current &&
i < bcd_reference_current->states_len) ?
bcd_reference_current->states[i] :
NULL;
/* See comment on `layer_type_is_dynamic` above. */
if (layer_type_is_dynamic) {
state_reference = NULL;
}
bcd->states[i] = BLI_array_store_state_add(
bs, layer->data, (size_t)data_len * stride, state_reference);
}
else {
bcd->states[i] = NULL;
}
}
if (layer->data) {
MEM_freeN(layer->data);
layer->data = NULL;
}
}
if (create) {
if (bcd_reference_current) {
bcd_reference_current = bcd_reference_current->next;
}
}
}
if (create) {
*r_bcd_first = bcd_first;
}
}
/**
* \note There is no room for data going out of sync here.
* The layers and the states are stored together so this can be kept working.
*/
static void um_arraystore_cd_expand(const BArrayCustomData *bcd,
struct CustomData *cdata,
const size_t data_len)
{
CustomDataLayer *layer = cdata->layers;
while (bcd) {
const int stride = CustomData_sizeof(bcd->type);
for (int i = 0; i < bcd->states_len; i++) {
BLI_assert(bcd->type == layer->type);
if (bcd->states[i]) {
size_t state_len;
layer->data = BLI_array_store_state_data_get_alloc(bcd->states[i], &state_len);
BLI_assert(stride * data_len == state_len);
UNUSED_VARS_NDEBUG(stride, data_len);
}
else {
layer->data = NULL;
}
layer++;
}
bcd = bcd->next;
}
}
static void um_arraystore_cd_free(BArrayCustomData *bcd)
{
while (bcd) {
BArrayCustomData *bcd_next = bcd->next;
const int stride = CustomData_sizeof(bcd->type);
BArrayStore *bs = BLI_array_store_at_size_get(&um_arraystore.bs_stride, stride);
for (int i = 0; i < bcd->states_len; i++) {
if (bcd->states[i]) {
BLI_array_store_state_remove(bs, bcd->states[i]);
}
}
MEM_freeN(bcd);
bcd = bcd_next;
}
}
/**
* \param create: When false, only free the arrays.
* This is done since when reading from an undo state, they must be temporarily expanded.
* then discarded afterwards, having this argument avoids having 2x code paths.
*/
static void um_arraystore_compact_ex(UndoMesh *um, const UndoMesh *um_ref, bool create)
{
Mesh *me = &um->me;
um_arraystore_cd_compact(
&me->vdata, me->totvert, create, um_ref ? um_ref->store.vdata : NULL, &um->store.vdata);
um_arraystore_cd_compact(
&me->edata, me->totedge, create, um_ref ? um_ref->store.edata : NULL, &um->store.edata);
um_arraystore_cd_compact(
&me->ldata, me->totloop, create, um_ref ? um_ref->store.ldata : NULL, &um->store.ldata);
um_arraystore_cd_compact(
&me->pdata, me->totpoly, create, um_ref ? um_ref->store.pdata : NULL, &um->store.pdata);
if (me->key && me->key->totkey) {
const size_t stride = me->key->elemsize;
BArrayStore *bs = create ? BLI_array_store_at_size_ensure(
&um_arraystore.bs_stride, stride, ARRAY_CHUNK_SIZE) :
NULL;
if (create) {
um->store.keyblocks = MEM_mallocN(me->key->totkey * sizeof(*um->store.keyblocks), __func__);
}
KeyBlock *keyblock = me->key->block.first;
for (int i = 0; i < me->key->totkey; i++, keyblock = keyblock->next) {
if (create) {
BArrayState *state_reference = (um_ref && um_ref->me.key && (i < um_ref->me.key->totkey)) ?
um_ref->store.keyblocks[i] :
NULL;
um->store.keyblocks[i] = BLI_array_store_state_add(
bs, keyblock->data, (size_t)keyblock->totelem * stride, state_reference);
}
if (keyblock->data) {
MEM_freeN(keyblock->data);
keyblock->data = NULL;
}
}
}
if (me->mselect && me->totselect) {
BLI_assert(create == (um->store.mselect == NULL));
if (create) {
BArrayState *state_reference = um_ref ? um_ref->store.mselect : NULL;
const size_t stride = sizeof(*me->mselect);
BArrayStore *bs = BLI_array_store_at_size_ensure(
&um_arraystore.bs_stride, stride, ARRAY_CHUNK_SIZE);
um->store.mselect = BLI_array_store_state_add(
bs, me->mselect, (size_t)me->totselect * stride, state_reference);
}
/* keep me->totselect for validation */
MEM_freeN(me->mselect);
me->mselect = NULL;
}
if (create) {
um_arraystore.users += 1;
}
}
/**
* Move data from allocated arrays to de-duplicated states and clear arrays.
*/
static void um_arraystore_compact(UndoMesh *um, const UndoMesh *um_ref)
{
um_arraystore_compact_ex(um, um_ref, true);
}
static void um_arraystore_compact_with_info(UndoMesh *um, const UndoMesh *um_ref)
{
# ifdef DEBUG_PRINT
size_t size_expanded_prev, size_compacted_prev;
BLI_array_store_at_size_calc_memory_usage(
&um_arraystore.bs_stride, &size_expanded_prev, &size_compacted_prev);
# endif
# ifdef DEBUG_TIME
TIMEIT_START(mesh_undo_compact);
# endif
um_arraystore_compact(um, um_ref);
# ifdef DEBUG_TIME
TIMEIT_END(mesh_undo_compact);
# endif
# ifdef DEBUG_PRINT
{
size_t size_expanded, size_compacted;
BLI_array_store_at_size_calc_memory_usage(
&um_arraystore.bs_stride, &size_expanded, &size_compacted);
const double percent_total = size_expanded ?
(((double)size_compacted / (double)size_expanded) * 100.0) :
-1.0;
size_t size_expanded_step = size_expanded - size_expanded_prev;
size_t size_compacted_step = size_compacted - size_compacted_prev;
const double percent_step = size_expanded_step ?
(((double)size_compacted_step / (double)size_expanded_step) *
100.0) :
-1.0;
printf("overall memory use: %.8f%% of expanded size\n", percent_total);
printf("step memory use: %.8f%% of expanded size\n", percent_step);
}
# endif
}
# ifdef USE_ARRAY_STORE_THREAD
struct UMArrayData {
UndoMesh *um;
const UndoMesh *um_ref; /* can be NULL */
};
static void um_arraystore_compact_cb(TaskPool *__restrict UNUSED(pool), void *taskdata)
{
struct UMArrayData *um_data = taskdata;
um_arraystore_compact_with_info(um_data->um, um_data->um_ref);
}
# endif /* USE_ARRAY_STORE_THREAD */
/**
* Remove data we only expanded for temporary use.
*/
static void um_arraystore_expand_clear(UndoMesh *um)
{
um_arraystore_compact_ex(um, NULL, false);
}
static void um_arraystore_expand(UndoMesh *um)
{
Mesh *me = &um->me;
um_arraystore_cd_expand(um->store.vdata, &me->vdata, me->totvert);
um_arraystore_cd_expand(um->store.edata, &me->edata, me->totedge);
um_arraystore_cd_expand(um->store.ldata, &me->ldata, me->totloop);
um_arraystore_cd_expand(um->store.pdata, &me->pdata, me->totpoly);
if (um->store.keyblocks) {
const size_t stride = me->key->elemsize;
KeyBlock *keyblock = me->key->block.first;
for (int i = 0; i < me->key->totkey; i++, keyblock = keyblock->next) {
BArrayState *state = um->store.keyblocks[i];
size_t state_len;
keyblock->data = BLI_array_store_state_data_get_alloc(state, &state_len);
BLI_assert(keyblock->totelem == (state_len / stride));
UNUSED_VARS_NDEBUG(stride);
}
}
if (um->store.mselect) {
const size_t stride = sizeof(*me->mselect);
BArrayState *state = um->store.mselect;
size_t state_len;
me->mselect = BLI_array_store_state_data_get_alloc(state, &state_len);
BLI_assert(me->totselect == (state_len / stride));
UNUSED_VARS_NDEBUG(stride);
}
}
static void um_arraystore_free(UndoMesh *um)
{
Mesh *me = &um->me;
um_arraystore_cd_free(um->store.vdata);
um_arraystore_cd_free(um->store.edata);
um_arraystore_cd_free(um->store.ldata);
um_arraystore_cd_free(um->store.pdata);
if (um->store.keyblocks) {
const size_t stride = me->key->elemsize;
BArrayStore *bs = BLI_array_store_at_size_get(&um_arraystore.bs_stride, stride);
for (int i = 0; i < me->key->totkey; i++) {
BArrayState *state = um->store.keyblocks[i];
BLI_array_store_state_remove(bs, state);
}
MEM_freeN(um->store.keyblocks);
um->store.keyblocks = NULL;
}
if (um->store.mselect) {
const size_t stride = sizeof(*me->mselect);
BArrayStore *bs = BLI_array_store_at_size_get(&um_arraystore.bs_stride, stride);
BArrayState *state = um->store.mselect;
BLI_array_store_state_remove(bs, state);
um->store.mselect = NULL;
}
um_arraystore.users -= 1;
BLI_assert(um_arraystore.users >= 0);
if (um_arraystore.users == 0) {
# ifdef DEBUG_PRINT
printf("mesh undo store: freeing all data!\n");
# endif
BLI_array_store_at_size_clear(&um_arraystore.bs_stride);
# ifdef USE_ARRAY_STORE_THREAD
BLI_task_pool_free(um_arraystore.task_pool);
um_arraystore.task_pool = NULL;
# endif
}
}
/** \} */
/* -------------------------------------------------------------------- */
/** \name Array Store Utilities
* \{ */
/**
* Create an array of #UndoMesh from `objects`.
*
* where each element in the resulting array is the most recently created
* undo-mesh for the object's mesh.
* When no undo-mesh can be found that array index is NULL.
*
* This is used for de-duplicating memory between undo steps,
* failure to find the undo step will store a full duplicate in memory.
* define `DEBUG_PRINT` to check memory is de-duplicating as expected.
*/
static UndoMesh **mesh_undostep_reference_elems_from_objects(Object **object, int object_len)
{
/* Map: `Mesh.id.session_uuid` -> `UndoMesh`. */
GHash *uuid_map = BLI_ghash_ptr_new_ex(__func__, object_len);
UndoMesh **um_references = MEM_callocN(sizeof(UndoMesh *) * object_len, __func__);
for (int i = 0; i < object_len; i++) {
const Mesh *me = object[i]->data;
BLI_ghash_insert(uuid_map, POINTER_FROM_INT(me->id.session_uuid), &um_references[i]);
}
int uuid_map_len = object_len;
/* Loop backwards over all previous mesh undo data until either:
* - All elements have been found (where `um_references` we'll have every element set).
* - There are no undo steps left to look for. */
UndoMesh *um_iter = um_arraystore.local_links.last;
while (um_iter && (uuid_map_len != 0)) {
UndoMesh **um_p;
if ((um_p = BLI_ghash_popkey(uuid_map, POINTER_FROM_INT(um_iter->me.id.session_uuid), NULL))) {
*um_p = um_iter;
uuid_map_len--;
}
um_iter = um_iter->local_prev;
}
BLI_assert(uuid_map_len == BLI_ghash_len(uuid_map));
BLI_ghash_free(uuid_map, NULL, NULL);
if (uuid_map_len == object_len) {
MEM_freeN(um_references);
um_references = NULL;
}
return um_references;
}
/** \} */
#endif /* USE_ARRAY_STORE */
/* for callbacks */
/* undo simply makes copies of a bmesh */
/**
* \param um_ref: The reference to use for de-duplicating memory between undo-steps.
*/
static void *undomesh_from_editmesh(UndoMesh *um, BMEditMesh *em, Key *key, UndoMesh *um_ref)
{
BLI_assert(BLI_array_is_zeroed(um, 1));
#ifdef USE_ARRAY_STORE_THREAD
/* changes this waits is low, but must have finished */
if (um_arraystore.task_pool) {
BLI_task_pool_work_and_wait(um_arraystore.task_pool);
}
#endif
/* make sure shape keys work */
if (key != NULL) {
um->me.key = (Key *)BKE_id_copy_ex(
NULL, &key->id, NULL, LIB_ID_COPY_LOCALIZE | LIB_ID_COPY_NO_ANIMDATA);
}
else {
um->me.key = NULL;
}
/* Uncomment for troubleshooting. */
// BM_mesh_validate(em->bm);
/* Copy the ID name characters to the mesh so code that depends on accessing the ID type can work
* on it. Necessary to use the attribute API. */
strcpy(um->me.id.name, "MEundomesh_from_editmesh");
BM_mesh_bm_to_me(
NULL,
em->bm,
&um->me,
(&(struct BMeshToMeshParams){
/* Undo code should not be manipulating 'G_MAIN->object' hooks/vertex-parent. */
.calc_object_remap = false,
.update_shapekey_indices = false,
.cd_mask_extra = {.vmask = CD_MASK_SHAPE_KEYINDEX},
.active_shapekey_to_mvert = true,
}));
um->selectmode = em->selectmode;
um->shapenr = em->bm->shapenr;
#ifdef USE_ARRAY_STORE
{
/* Add ourselves. */
BLI_addtail(&um_arraystore.local_links, um);
# ifdef USE_ARRAY_STORE_THREAD
if (um_arraystore.task_pool == NULL) {
um_arraystore.task_pool = BLI_task_pool_create_background(NULL, TASK_PRIORITY_LOW);
}
struct UMArrayData *um_data = MEM_mallocN(sizeof(*um_data), __func__);
um_data->um = um;
um_data->um_ref = um_ref;
BLI_task_pool_push(um_arraystore.task_pool, um_arraystore_compact_cb, um_data, true, NULL);
# else
um_arraystore_compact_with_info(um, um_ref);
# endif
}
#else
UNUSED_VARS(um_ref);
#endif
return um;
}
static void undomesh_to_editmesh(UndoMesh *um, Object *ob, BMEditMesh *em)
{
BMEditMesh *em_tmp;
BMesh *bm;
#ifdef USE_ARRAY_STORE
# ifdef USE_ARRAY_STORE_THREAD
/* changes this waits is low, but must have finished */
BLI_task_pool_work_and_wait(um_arraystore.task_pool);
# endif
# ifdef DEBUG_TIME
TIMEIT_START(mesh_undo_expand);
# endif
um_arraystore_expand(um);
# ifdef DEBUG_TIME
TIMEIT_END(mesh_undo_expand);
# endif
#endif /* USE_ARRAY_STORE */
const BMAllocTemplate allocsize = BMALLOC_TEMPLATE_FROM_ME(&um->me);
em->bm->shapenr = um->shapenr;
EDBM_mesh_free_data(em);
bm = BM_mesh_create(&allocsize,
&((struct BMeshCreateParams){
.use_toolflags = true,
}));
BM_mesh_bm_from_me(bm,
&um->me,
(&(struct BMeshFromMeshParams){
/* Handled with tessellation. */
.calc_face_normal = false,
.calc_vert_normal = false,
.active_shapekey = um->shapenr,
}));
em_tmp = BKE_editmesh_create(bm);
*em = *em_tmp;
/* Normals should not be stored in the undo mesh, so recalculate them. The edit
* mesh is expected to have valid normals and there is no tracked dirty state. */
BLI_assert(BKE_mesh_vertex_normals_are_dirty(&um->me));
/* Calculate face normals and tessellation at once since it's multi-threaded. */
BKE_editmesh_looptri_and_normals_calc(em);
em->selectmode = um->selectmode;
bm->selectmode = um->selectmode;
bm->spacearr_dirty = BM_SPACEARR_DIRTY_ALL;
ob->shapenr = um->shapenr;
MEM_freeN(em_tmp);
#ifdef USE_ARRAY_STORE
um_arraystore_expand_clear(um);
#endif
}
static void undomesh_free_data(UndoMesh *um)
{
Mesh *me = &um->me;
#ifdef USE_ARRAY_STORE
# ifdef USE_ARRAY_STORE_THREAD
/* changes this waits is low, but must have finished */
BLI_task_pool_work_and_wait(um_arraystore.task_pool);
# endif
/* we need to expand so any allocations in custom-data are freed with the mesh */
um_arraystore_expand(um);
BLI_assert(BLI_findindex(&um_arraystore.local_links, um) != -1);
BLI_remlink(&um_arraystore.local_links, um);
um_arraystore_free(um);
#endif
if (me->key) {
BKE_key_free_data(me->key);
MEM_freeN(me->key);
}
BKE_mesh_free_data_for_undo(me);
}
static Object *editmesh_object_from_context(bContext *C)
{
Scene *scene = CTX_data_scene(C);
ViewLayer *view_layer = CTX_data_view_layer(C);
BKE_view_layer_synced_ensure(scene, view_layer);
Object *obedit = BKE_view_layer_edit_object_get(view_layer);
if (obedit && obedit->type == OB_MESH) {
Mesh *me = obedit->data;
if (me->edit_mesh != NULL) {
return obedit;
}
}
return NULL;
}
/** \} */
/* -------------------------------------------------------------------- */
/** \name Implements ED Undo System
*
* \note This is similar for all edit-mode types.
* \{ */
typedef struct MeshUndoStep_Elem {
UndoRefID_Object obedit_ref;
UndoMesh data;
} MeshUndoStep_Elem;
typedef struct MeshUndoStep {
UndoStep step;
MeshUndoStep_Elem *elems;
uint elems_len;
} MeshUndoStep;
static bool mesh_undosys_poll(bContext *C)
{
return editmesh_object_from_context(C) != NULL;
}
static bool mesh_undosys_step_encode(struct bContext *C, struct Main *bmain, UndoStep *us_p)
{
MeshUndoStep *us = (MeshUndoStep *)us_p;
/* Important not to use the 3D view when getting objects because all objects
* outside of this list will be moved out of edit-mode when reading back undo steps. */
const Scene *scene = CTX_data_scene(C);
ViewLayer *view_layer = CTX_data_view_layer(C);
ToolSettings *ts = CTX_data_tool_settings(C);
uint objects_len = 0;
Object **objects = ED_undo_editmode_objects_from_view_layer(scene, view_layer, &objects_len);
us->elems = MEM_callocN(sizeof(*us->elems) * objects_len, __func__);
us->elems_len = objects_len;
UndoMesh **um_references = NULL;
#ifdef USE_ARRAY_STORE
um_references = mesh_undostep_reference_elems_from_objects(objects, objects_len);
#endif
for (uint i = 0; i < objects_len; i++) {
Object *ob = objects[i];
MeshUndoStep_Elem *elem = &us->elems[i];
elem->obedit_ref.ptr = ob;
Mesh *me = elem->obedit_ref.ptr->data;
BMEditMesh *em = me->edit_mesh;
undomesh_from_editmesh(
&elem->data, me->edit_mesh, me->key, um_references ? um_references[i] : NULL);
em->needs_flush_to_id = 1;
us->step.data_size += elem->data.undo_size;
elem->data.uv_selectmode = ts->uv_selectmode;
#ifdef USE_ARRAY_STORE
/** As this is only data storage it is safe to set the session ID here. */
elem->data.me.id.session_uuid = me->id.session_uuid;
#endif
}
MEM_freeN(objects);
if (um_references != NULL) {
MEM_freeN(um_references);
}
bmain->is_memfile_undo_flush_needed = true;
return true;
}
static void mesh_undosys_step_decode(struct bContext *C,
struct Main *bmain,
UndoStep *us_p,
const eUndoStepDir UNUSED(dir),
bool UNUSED(is_final))
{
MeshUndoStep *us = (MeshUndoStep *)us_p;
ED_undo_object_editmode_restore_helper(
C, &us->elems[0].obedit_ref.ptr, us->elems_len, sizeof(*us->elems));
BLI_assert(BKE_object_is_in_editmode(us->elems[0].obedit_ref.ptr));
for (uint i = 0; i < us->elems_len; i++) {
MeshUndoStep_Elem *elem = &us->elems[i];
Object *obedit = elem->obedit_ref.ptr;
Mesh *me = obedit->data;
if (me->edit_mesh == NULL) {
/* Should never fail, may not crash but can give odd behavior. */
CLOG_ERROR(&LOG,
"name='%s', failed to enter edit-mode for object '%s', undo state invalid",
us_p->name,
obedit->id.name);
continue;
}
BMEditMesh *em = me->edit_mesh;
undomesh_to_editmesh(&elem->data, obedit, em);
em->needs_flush_to_id = 1;
DEG_id_tag_update(&me->id, ID_RECALC_GEOMETRY);
}
/* The first element is always active */
ED_undo_object_set_active_or_warn(
CTX_data_scene(C), CTX_data_view_layer(C), us->elems[0].obedit_ref.ptr, us_p->name, &LOG);
/* Check after setting active. */
BLI_assert(mesh_undosys_poll(C));
Scene *scene = CTX_data_scene(C);
scene->toolsettings->selectmode = us->elems[0].data.selectmode;
scene->toolsettings->uv_selectmode = us->elems[0].data.uv_selectmode;
bmain->is_memfile_undo_flush_needed = true;
WM_event_add_notifier(C, NC_GEOM | ND_DATA, NULL);
}
static void mesh_undosys_step_free(UndoStep *us_p)
{
MeshUndoStep *us = (MeshUndoStep *)us_p;
for (uint i = 0; i < us->elems_len; i++) {
MeshUndoStep_Elem *elem = &us->elems[i];
undomesh_free_data(&elem->data);
}
MEM_freeN(us->elems);
}
static void mesh_undosys_foreach_ID_ref(UndoStep *us_p,
UndoTypeForEachIDRefFn foreach_ID_ref_fn,
void *user_data)
{
MeshUndoStep *us = (MeshUndoStep *)us_p;
for (uint i = 0; i < us->elems_len; i++) {
MeshUndoStep_Elem *elem = &us->elems[i];
foreach_ID_ref_fn(user_data, ((UndoRefID *)&elem->obedit_ref));
}
}
void ED_mesh_undosys_type(UndoType *ut)
{
ut->name = "Edit Mesh";
ut->poll = mesh_undosys_poll;
ut->step_encode = mesh_undosys_step_encode;
ut->step_decode = mesh_undosys_step_decode;
ut->step_free = mesh_undosys_step_free;
ut->step_foreach_ID_ref = mesh_undosys_foreach_ID_ref;
ut->flags = UNDOTYPE_FLAG_NEED_CONTEXT_FOR_ENCODE;
ut->step_size = sizeof(MeshUndoStep);
}
/** \} */